Why Google's Street View WiFi Data Collection Was Almost Certainly An Accident

from the technical-details dept

We’ve been among those who have believed that Google’s collection of WiFi data via its Street View cars was likely an accident — but some have argued that it is impossible to do such a thing by accident. In fact, in the various lawsuits and legal maneuverings around this mess, many people keep claiming that there’s simply no way Google was accidentally collecting this data — although we’ve yet to hear a single person explain what Google would possibly want with the data, or seen a single shred of evidence that anything was ever done with the data. However, for those who insist it is impossible to for this to have happened by accident, Slashdot points us to a detailed technical analysis of why it almost certainly was an accident, despite all the claims to the contrary.

It explains, in great detail, how and why the collection of data packets would occur, mainly to help triangulate where the WiFi network was located — something that Google has always admitted to doing. The problem was that some of the junk data (a very tiny amount, again, as explained in the article) got caught and retained, when it should have been dumped:

Although some people are suspicious of their explanation, Google is almost certainly telling the truth when it claims it was an accident. The technology for WiFi scanning means it’s easy to inadvertently capture too much information, and be unaware of it.

It then goes on to show how all of this works, using a specific example from within a Panera Bread restaurant that has open WiFi, which the author uses to demonstrate just how easy it is to capture stray data, why it would make sense and also just how useless most of that data really would be. It’s pretty convincing, but I doubt it will satisfy the conspiracy theorists who are just absolutely positive Google had something nefarious planned.

The key issue, as has been pointed out repeatedly, is that most people arguing nefarious intent don’t seem to understand what Google was actually doing. It was trying to map the location of WiFi base-stations, a perfectly legal activity that a small group of companies have been doing for years. But in order to best figure out the location of the networks, it’s helpful to have as much data as possible that traversing over the access point. The system doesn’t care or need to know what that data is, it just wants as much data as possible for the purpose of triangulating. The problem was that Google’s system “kept” the data that it got, even though there’s been no evidence presented that the the data was ever used for anything (a key point that those screaming “criminal intent” repeatedly gloss over). On top of that, no one even explains why Google would want such data. The little snippets would be so random it’s difficult to come up with any reason why keeping such data would be useful.

Triangulation is a lot harder than you’d think. This is because many things will block or reflect the signal. Therefore, as the car drives buy, it wants to get every single packet transmitted by the access-point in order to figure out its location. Curiously, with all that data, Google can probably also figure out the structure of the building, by finding things like support columns that obstruct the signal.

What’s important about this packet is that Google only cares about the MAC addresses found in the header, and the signal strength, but doesn’t care about the payload. If you look further down in the payload [in the example data from an open WiFi network in Panera], you’ll notice that it’s inadvertently captured a URL.

Take a look again. Even though the access-point MAC address is highlighted, there’s extra data in the packet. These extra data will include URLs, fragments of data returned from websites (like images), the occasional password, cookies, fragments of e-mails, and so on. However, the quantity of this information will be low compared to the total number of packets sniffed by Google.

That’s the core of this problem. Google sniffed packets, only caring about MAC addresses and SSIDs, but when somebody did an audit, they found that the captured packets occasionally contained more data, such as URLs and e-mail fragments.

I agree with the conclusion to the post. Just because this was pretty clearly an accident, it still doesn’t make it a good thing. Google clearly should have realized this much earlier and never allowed such data to be captured. But those running around screaming about how this was all pre-meditated by Google are going to have to offer up a lot more evidence.

Filed Under: , ,

Rate this comment as insightful
Rate this comment as funny
You have rated this comment as insightful
You have rated this comment as funny
Flag this comment as abusive/trolling/spam
You have flagged this comment
The first word has already been claimed
The last word has already been claimed
Insightful Lightbulb icon Funny Laughing icon Abusive/trolling/spam Flag icon Insightful badge Lightbulb icon Funny badge Laughing icon Comments icon

Comments on “Why Google's Street View WiFi Data Collection Was Almost Certainly An Accident”

Subscribe: RSS Leave a comment
123 Comments
Shawn (profile) says:

Google is guilty of poor quality control on the WiFi mapping portion of it’s stretview project. That poor QC caused them to unintentionally write data to disk that they did not require to complete the project as designed. That is about it. I have seen no evidence of any conspiracy to collect the data or do ANYTHING with the data they unintentionally collected.

Hephaestus (profile) says:

Re: Re:

Google shill !!! Someone check this guys IP adddress and find out if he is working for google … Big Ole GRIN!!

“Google is guilty of poor quality control on the WiFi mapping portion of it’s stretview project. That poor QC caused them to unintentionally write data to disk that they did not require to complete the project as designed. “

One of the projects I am working on does something similar because of a closed source driver. Sometimes things like this are unavoidable and you dont realize its going on until much later.

out_of_the_blue says:

Curious form of argument: state possible reasons,

while denying that there *is* reason for concern. You’ve listed far more nefarious reasons than I could have come up with. I particularly like: “can probably also figure out the structure of the building”.

You end merely stating that we’ve no evidence of *actual* mis-use. That’s what the process of discovery is for; we should be able to find out what if any purpose they have. — Surely our wonderful court system will soon find them not guilty, right? — But I think that the average jury will take a dim view of this “incidental” snooping; the whole Street View scheme verges on invasion of privacy.

Christopher Weigel (profile) says:

Re: Curious form of argument: state possible reasons,

Oof. The point was that their objective (finding the wi-fi access point) is harder than just grabbing a couple packets due to obstruction. The more data the better the determined location, and as a consequence you can find where there are obstructions.

Are you *really* so stupid and/or desperate to find something google “did wrong” as to have just accused them of trying to find structural weaknesses in buildings?

The data capture was a simple side-effect of their intended goal, i.e. to capture MAC addresses and signal strength. An unfortunate side effect, which is way overblown by idiots like, well, you.

out_of_the_blue says:

Re: Re: @Christopher Weigel

Thank you for your remarks. The line I quoted is in the quoted section of the piece above, and I think my taking it as amusing is obvious enough.

I’ve skimmed your profile. You’re merely negativism, usually a “drive-by” one-line ad hominem attack, delivered with condescension and the occasional “STFU”, and evidence a complete lack of self-awareness that you are what you project onto others: a common troll. You’ll remain so until you offer opinions.

Berenerd (profile) says:

Re: Curious form of argument: state possible reasons,

The general specifics such as “OHNOES THERE IS A CONCRETE ALL BETWEEN US AN THE AP!!!!11111oneoneoneone!” Not “Oh, its in the Green house with the kids playing craps outfront!” Obstructions don’t even need to be the building its in but a building between you and it.

Seth says:

I agree.

I still don’t get why everyone is making such a fuss about it. The data collected all originated from unencrypted networks anybody can intercept. Also, no evident malicious intent was shown.

Maybe it’s not nice that they captured it, but instead of analyzing the background of the problem, everyone is jumping on big G.

I’m generally very cautious about my data, but this is just a farce, showing how people search someone to blame instead of fixing the damn problem.

Eric says:

First of all, your link is to an outdated speculative technical analysis. Try this for a fact-based analysis of what Google was actually doing. You can see that in the source code for gslite (Google’s collection software), it specifically discards encrypted data while writing unencrypted data to disk. It is not reasonable to argue that Google “accidentally” coded their software this way. It is also tenuous to argue that this code somehow slipped through the cracks, since Google itself claims that it “makes extensive use of code reviews to ensure quality in its products; every line of Google source code is reviewed in this way.” (emphasis added)

Also, I point your attention to Skyhook, which uses much less intrusive “active” scanning while still gathering the only two pieces of information needed to provide their location service: MAC addresses and signal strength. Skyhook’s method makes any such payload data gathering impossible.

One final point. Whether or not Google’s board of directors decided they wanted to gather payload data (which I do not believe to be the case), the words “intent” and “accident” may mean different things when put in a legal context.

Modplan (profile) says:

Re: Re:

1) Encrypted data is fairly useless any way.

2) Just because Google claims they have high QA standards does not suddenly mean they never have problems. Especially in situations like this where the type and amount of data you can get is effectively useless for nefarious purposes, which makes that particular change less important. It’s a (small) configuration issue, not badly written code.

3) Skyhook does not appear to be aimed at the same situation that google was in, where cars were passing by and were far more pressed in the time and information they could get at any one place.

Eric says:

Re: Re: Re:

1) What’s your point? I’m not suggesting that Google should have stored encrypted data. I’m saying that the fact that they discarded encrypted payloads while keeping unencrypted payloads is evidence of intent.

2) I didn’t say the code was badly written. In fact, I would say it’s well-written. Well-written + code review = Just more evidence of intent.

3) I’m not totally sure what your point is here… In any case, Skyhook provides the exact same location service, but collects FAR less information in a much less intrusive way.

Modplan (profile) says:

Re: Re: Re: Re:

2) No it doesn’t. Code review != configuration review. Not to mention you’re referring to what is effectively marketing material – I highly doubt any claim that every individual line is reviewed, especially to the degree that this particular configuration would be caught.

3) My point is that the skyhook example was irrelevant. There was no detail in that link at all as to how Skyhook gains its database of Wi-Fi hot spots. All it was an advert for a location service/technology that happened to use Wi-Fi hot spot information, nothing about the specifics of their data collection software, their policy regarding retention of data gained from unencrypted networks as they gather hot spot info, or anything else that might be of use.

Skyhook provides a location service, but I have no idea as to the methods they use to gather hot spot info nor whether their software retains unencrypted info. In fact the real link to info about their Wi-FI information gathering is here:

http://www.skyhookwireless.com/howitworks/coverage.php

But again, carries no real information. It’s general promotional material, not specific details.

Practically none of what you have presented goes against Google.

Eric says:

Re: Re: Re:2 Re:

2) Legally, yes it does. You can’t configure code to do something the code was not designed to do in the first place. That’s the issue. Google intended to write code that could (AND DID) collect all unencrypted traffic on open networks. This is a fact, and is extremely convincing evidence of intent. Whether or not they meant to configure it another way is a different question, but the fact that they ran it in the data-scarfing configuration is more evidence of intent. And marketing material or not, it’s an official statement from Google and it WILL be used against them in court.

3) Here’s a better link for how Skyhook works: http://www.xconomy.com/boston/2010/05/17/googles-passive-sniffing-technique-may-have-paved-the-way-f or-wi-fi-privacy-flap-skyhook

Skyhook’s architecture does not, in fact CANNOT, see or capture payload data. The only data Skyhook gets is beacon packets from access points. Google sniffed data from access points AND clients.

Modplan (profile) says:

Re: Re: Re:3 Re:

This is a fact, and is extremely convincing evidence of inten

No it isn’t. It’s evidence that the software has multiple ways it can be configured and has multiple uses, like any other piece of software. It’s evidence that someone configured it to do something that lead to this PR disaster, but it says nothing about intent. The fact that you regularly make claims to intent, and not the facts of what actually happened, including the specific information gathered and whether it could be in any way useful considering such short periods of data collection indicates how much of an agenda you have in pursuing this.

In fact, the history of the program used even indicates the issue is in configuration, not development/code review.

3) Skyhook describes 2 different methods of data collection (already addressed by another poster further down), that’s about it. This doesn’t show anything about intent, in fact Skyhook itself notes the issue with using the other method, being you miss access points if they’re busy (which is likely for public Wi-Fi spots).

From the article:

Nor would it be possible for Google to record such data completely by accident, Morgan says. “At the engineering level it’s very easy to know whether you are capturing this data or not,” he says. So the error at Google, he says, probably happened “higher up the food chain…An engineer doesn’t care, and grabs whatever he can. But when there’s no one looking at it who’s got the broader perspective to understand the implications, that’s where the breakdown happens.”

Even Skyhook understands that there’s a difference between coed review and how the software is setup.

eric says:

Re: Re: Re:4 Re:

Do you disagree with this?: The fact that gslite’s default behavior is to record unencrypted payload data is evidence of intent to create a program that has the default behavior to record unencrypted payload data.

Of course you don’t.

Also, believe it or not, the fact that Google did not choose to execute the program with different command line arguments is also legally evidence that they intended to run it that way. Now, there may be rebutting evidence. But right now, the only rebutting “evidence” we have is what Google says. You’re welcome to just blindly accept what Google says (or rather, what Google’s PR team approves).

Modplan (profile) says:

Re: Re: Re:5 Re:

The only evidence you have of Google doing anything illegal is based on misreadings of the law about having to do purely with intent to record data and relying on the claim that they intended to use publicly accessible data, which no one denies.

Intentionally is a clarification to differentiate between unintentional (someone who doesn’t know and/or realise what they’re doing or that they’ve done so), not that intent in itself is illegal, especially when you actually take into account the various exemptions to wiretapping not being illegal (or not even being considered wiretapping by popular use of the term), which includes when information is broadcast over publicly accessible communications like unecrypted wi-fi, and assuming you haven’t used the information for nefarious purposes, which there is no evidence of.

You’re welcome to blindly parrot conspiracy theories that Google is genuinely wiretapping to use data for nefarious, illegal purposes based on misreadings of the law and misunderstandings and misuse of technology and technological terms.

eric says:

Re: Re: Re:6 Re:

I don’t claim that Google illegally intercepted publicly accessible information. Such a claim would be silly, because it’s entirely legal to intercept publicly accessible information. I’m arguing that what Google collected wasn’t publicly accessible at all. You’re welcome to disagree on that point, but it’s not you who gets to decide. A district court judge in Northern California will decide that.

I agree with your understanding of “intentionally” — but you’re relying solely on Google’s word for the assertion that they didn’t “know and/or realize what they’re doing or that they’ve done so.” Instead of relying on what Google has said, I’m looking at the entire body of evidence, and to me it’s plain that their activity was intentional. Again, you’re welcome to disagree on this point. But again, a district court judge in California will get to decide.

Please quote my “conspiracy theories.” I think if you re-read what I’ve written, you’ll see that I never claimed Google did anything at all with the data they captured (besides wanting desperately to delete it). I do argue that they potentially violated the Wiretap Act (which, as you should know by now, does not require malicious or nefarious intent, nor does it require anyone to suffer harm, nor does it require Google do actually do anything with the data).

Tek'a R (profile) says:

Re: Re: Re:7 Re:

I’m arguing that what Google collected wasn’t publicly accessible at all.

an unencrypted wireless hotspot, whether in your home or in a coffee shop, is Incredibly accessible. All of the data moving across and through is likewise accessible.

This is not a case of peeking through windows or trying to lean close enough to hear someones phone conversation. This is making a note of where you saw the lit billboard with BELKINWIRELESS ROUTERHOME written across it, or where you heard an announcement across a PA system “WIRELESS CONNECTION ON AISLE TWELVE! PLEASE TAKE INFORMATION FROM THIS WIRELESS CONNECTION”

Starting to whimper and moan about Wiretapping legislation and claiming the entire body of evidence shows “intentional” activity is just silly.

Anonymous Coward says:

Re: Re: Re:7 Re:

Remember the case of the pedo that got caught because a police officer looked inside his folders through an opening in his system?

The judge in that case said if they guy didn’t protect his data and opened even by accident then the information was public.

So if the courts want to go that way Google is obviously clear at least in the U.S. and it will have no serious repercussion, on the other hand if they rule Google is responsible then they also need to re evaluate the decision for that pedo guy.

Now which one will the judges choose? Loose part of the ability to snoop onto others or protect the privacy of the population?

darryl says:

Re: Re: Re:2 Re:

“2) No it doesn’t. Code review != configuration review. “

What ???? YES, it does, and unit testing, field trials, usability, and suitability trials, “legals”.

Code review in one part of a complate picture, are you saying you can review how code operates, without considering its configuration ?

Do much coding yourself ?

Someone also said, “engineers dont write laws” true, but they must abide by the laws, and professional engineers know about the laws regarding their field of expertise.

You think an engineer who deals with electrical power would not be aware of the laws regarding safe wiring, and correct legal procedures.

Its the first thing engineers learn,

Someone also said, “it’s just because they use libraries”, yes they do, but they do not HAVE to use the library functions just because they are in the library, if a library function gathers unwanted, or illegal data, you dont call that function.

NO, this was a premeditated decision from management, the crews in the fields did exactly what they were instructed to do, it was no accident. It was planned, they got caught and now they are acting stupid to try to mitigate what they have done.

Stupid, in saying it was an accident, It just happened, in multiple countries, and some of the projects lasted for 3 years !!. And during that 3 years they never reviewed the collected data ?

NO, they knew what they were doing, got caught, adding to just another Google SNAFU.

rl78 says:

Re: Re: Re:3 your insights on this

I tend to agree with you.

I’m not so informed on wireless data transmissions or software dev to form my own decisive opinion on what Google did here but my logic led me to believe this:

http://techdirt.com/article.php?sid=20100622/0340389918#c1447

Just from reading your perspective, I felt your knowledgeable enough to say what you have said so I want to ask you do you agree with my assessment. Its from a logical general standpoint. I do have experience in IT, PC Tech support and general networking, but like I said not so much with programming and wireless data trans. I have seen what I felt were knowledgeable opinions defending Google, but still feel they were aware of what they were doing.

rl78 says:

Re: Re: Re:3 your insights on this

I tend to agree with you.
I’m not so informed on wireless data transmissions or software dev to form my own decisive opinion on what Google did here but my logic led me to believe this:
http://techdirt.com/article.php?sid=20100622/0340389918#c1447
Just from reading your perspective, I felt your knowledgeable enough to say what you have said so I want to ask you do you agree with my assessment. Its from a logical general standpoint. I do have experience in IT, PC Tech support and general networking, but like I said not so much with programming and wireless data trans. I have seen what I felt were knowledgeable opinions defending Google, but still feel they were aware of what they were doing.

greenbird (profile) says:

Re: Re: Re: Intent of what exactly?

Intent of what exactly? Intent of serendipitously saving information that people were openly broadcasting over the radio? This has got to be one of the most ridiculous witch hunts in a while. And it’s propagated by people like you claiming to understand and using words like “intent” to sound all legal and nefarious to scare people who have an even less understanding.

If I were writing the software I would have written it to store potentially useful data also by default as long as it cause no issues with the primary function of the software. It never hurts to have too much data while not having enough is unfixable. If you don’t need it you throw it away.

The data was placed out there, by the owner of the data, publicly available to anyone with a wireless network card. The only way intent is relevant from criminal perspective is if the intent of the data capture was to use it for illegal purposes. Good luck with that.

This is like suing Google because someone found they could read the note they put on their front door in the picture Google took (I shouldn’t mention it. It’ll be the headline tomorrow). You can’t arrest someone for reading a note you put on your door where anyone can read it even if you’re stupid enough to put your credit card number or SSN on it. Now if someone uses it illegally it’s a bit different. The key is that the collection of the information was in no way an issue either criminally or from a tort perspective if you placed it out in public. Especially in the case of someone driving by randomly collecting it. To claim otherwise shows a clear ignorance of what happened. This explains the political grandstanding about it.

MD (profile) says:

Re: Re: Re:2 Intent of what exactly?

You know, the way US privacy laws are written you can only collect if the information obtained is proven to cause harm/damage.

“Haha I have your SSN, birth date, driver’s license, mother’s maiden name, etc. and as long as I don’t use it for illegal purposes I can’t get in trouble.”

eric says:

Re: Re: Re:2 Intent of what exactly?

The question of intent goes to whether Google “intentionally intercepted” the traffic. This point was not chosen by me to create fear; it’s in the text of the wiretap act (18 U.S.C. §2511). (And I actually am a law clerk, by the way.)

This point I think cuts right to the heart of it:

If I were writing the software I would have written it to store potentially useful data also by default as long as it cause no issues with the primary function of the software. It never hurts to have too much data while not having enough is unfixable. If you don’t need it you throw it away.

Yes, from an engineering perspective, of course you save that data! It could be useful some day, and who wants to go back into the code to add it later? Why not just put that functionality in there the first time?

But engineers don’t write laws, legislators do. And the Wiretap Act is very clear.

It remains to be seen whether a court will find that the payload data was “readily accessible by the general public” (which is the language in the Wiretap Act). People who are very familiar with the technology seem to think that the data is “readily accessible to the general public.” But techies don’t make those decisions; courts do. And I personally think the courts will rule that it isn’t.

Modplan (profile) says:

Re: Re: Re:3 Intent of what exactly?

Intentionally intercepting unecrypted, open data is not illegal unless you use personal information for illegal purposes.

(c) It shall not be unlawful under this chapter for a person acting under color of law to intercept a wire, oral, or electronic communication, where such person is a party to the communication or one of the parties to the communication has given prior consent to such interception.

(d) It shall not be unlawful under this chapter for a person not acting under color of law to intercept a wire, oral, or electronic communication where such person is a party to the communication or where one of the parties to the communication has given prior consent to such interception unless such communication is intercepted for the purpose of committing any criminal or tortious act in violation of the Constitution or laws of the United States or of any State.

(g) It shall not be unlawful under this chapter or chapter 121 of this title for any person—
(i) to intercept or access an electronic communication made through an electronic communication system that is configured so that such electronic communication is readily accessible to the general public;

Note you misleading statement of the law.

eric says:

Re: Re: Re:4 Intent of what exactly?

Note that you didn’t read my whole post.

I’ll copy/paste the last paragraph again, which CLEARLY addresses the section you have in bold:

It remains to be seen whether a court will find that the payload data was “readily accessible [to] the general public” (which is the language in the Wiretap Act). People who are very familiar with the technology seem to think that the data is “readily accessible to the general public.” But techies don’t make those decisions; courts do. And I personally think the courts will rule that it isn’t.

eric says:

Re: Re: Re:6 Intent of what exactly?

Just because you think unencrypted Wi-Fi is “readily accessible to the general public” doesn’t make it legally true.

In fact I addressed this very point, so I’ll quote myself once more:

People who are very familiar with the technology seem to think that the data is “readily accessible to the general public.” But techies don’t make those decisions; courts do. And I personally think the courts will rule that it isn’t.

eric says:

Re: Re: Re:8 Intent of what exactly?

Well you’re right that my opinion about the case is not a fact. And have not said, “Google violated the Wiretap Act and that’s a fact!” I’ve been very careful not to say that. So if you’re interpreting my argument as saying that, then you’re misinterpreting.

But my opinion is based on facts. And yours is based on ad hominem, “personal agenda” or “conspiracy theory” attacks.

eric says:

Re: Re: Re:8 Intent of what exactly?

Yes there are other privacy laws. But it doesn’t matter if they didn’t violate other privacy laws. The only thing that matters is whether or not Google violated the Wiretap Act. And I think you’ll agree if that if a court finds (1) Google “intentionally intercepted” data that was (2) not “readily accessible to the general public,” they were in violation.

My point was that under the Wiretap Act, there is NO REQUIREMENT WHATSOEVER that Google do ANYTHING with the data they intercepted besides write it to disk; they are liable if the above two points are satisfied.

Joshua Jones (profile) says:

Re: Re: Re:9 Intent of what exactly?

And I think you’ll agree if that if a court finds (1) Google “intentionally intercepted” data that was (2) not “readily accessible to the general public,” they were in violation.

So… to rephrase..

And I think you’ll agree if a court finds them guilty, they were guilty

And for the record, accepting your apology for confusing beacon packets and SSIDs, doesn’t make me “see that [you] actually do have a very good understanding of the technology and what happened.

The whole problem with this situation is that people are flipping out even though:

1. The information collected is virtually useless, given the manner in which it was collected.

2. The information was pulled from open networks. Whether there were some networks that weren’t broadcasting their SSID seems irrelevant to me. If you know how to setup your network to not broadcast the SSID and still run fine, then you know how to turn on WEP or WPA encryption. If it’s open, it’s open. End of story.

Separately, am I the only one that has a weird feeling about constantly comparing Google in this instance to Skyhook, and the linking to what are basically promotional materials? For someone claiming that people are simply following along with Google’s PR, you kind of come off sounding like Skyhook PR. “Google doesn’t care about your privacy, but we do!

Free Capitalist (profile) says:

Re: Re: Re:3 Intent of what exactly?

The question of intent goes to whether Google “intentionally intercepted” the traffic.

I fail to see how the wiretap act could ever be applied to data which was broadcast in the clear over a radio frequency.

Radio waves are indiscriminate. Should I be arrested for tuning in to NPR during my morning commute? Well, yes, actually but that’s another story.

Users of wireless networks are broadcasting their data over radio, encrypted or not. Unencrypted network data is no different that unencrypted analog radio… anyone with an antenna and tuner in range has every right observe what waves they are being bombarded with.

Of course if Google picked up and stored the session data of someone downloading copyright infringing material.. well then they’re fucked.

Anonymous Coward says:

Re: Re:

Saving the packet bodies was configurable (could be enabled or disabled), probably for debugging. They probably just forgot to disable that setting before going into production.

And active scanning being less intrusive? I would expect the opposite: passively listening for the beacons is much less intrusive than actively sending probe requests and waiting for the responses (which uses bandwidth which could instead be used for data packets).

Anonymous Coward says:

Re: Re: Re: Re:

Eric,
The link has no discussion. http://www.my80211.com/home/2010/1/11/80211-client-active-and-passive-scanning.html would provide a better difference between the two. The issue with the code is that no matter how many code reviews or QA tests you run you will never catch all the bugs in a system. Intent would be if there was code that extracted the payload data specifically.

Anonymous Coward says:

Re: Re: Re:3 Re:

You do realize that if a wireless access point is not broadcasting beacon packets then it is impossible for a client to connect to it?

I I can tell you are trying to say networks that are not broadcasting the SSID, but it is cute to watch you condescend people who do not understand what you think you understand 😉

Anonymous Coward says:

Re: Re: Re:4 Re:

Well, being nitpicky, not impossible to connect (the client could do all the probe/authenticate dance without ever seeing a single beacon).

It is just that, without beacons, it would quickly disconnect (as in, within a few seconds or less) due to beacon loss (it would look to the client as if either it had moved a long distance from the access point in a matter of milliseconds, or the access point lost its power). The end effect to the user would be about the same (impossible to use the access point).

Eric says:

Re: Re: Re:4 Re:

I do know they can’t connect, but they don’t need to connect to peek at unencrypted header information. Google doesn’t need an IP address or a default gateway, they just need to intercept the wireless transmissions. That can be done without “connecting” to anything. In fact, Google never once “connected” (as in, actually tried to get an IP address) to any of the networks it snooped on.

Anonymous Coward says:

Re: Re: Re:5 Re:

No one is claiming Google tried to connect to anything.
You are thumping your chest saying you know as much about the topic as anyone outside of Google then claim they have the ability to collect data from access points that are not broadcasting! (impossible)
no wait! not broadcasting beacon packets ( A configuration that while is technically possible would not be found in the real world as the network would be unusable)

eric says:

Re: Re: Re:6 Re:

Are you serious? You realize you can configure a hotspot such that it does not broadcast its SSID? That is, when you’re in range, your typical wireless card interface will not give you an option to connect. And also, did you know that it’s possible to connect to a hotspot that is not broadcasting its SSID? In fact, I have configured my own router to do this exact thing.

I did a very quick Google search and found: http://compnetworking.about.com/cs/wirelessproducts/qt/disablessidcast.htm

Perhaps I have been sloppy with my language, but when I say Google snooped on networks that weren’t broadcasting, these are the networks I’m talking about.

Anonymous Coward says:

Re: Re: Re:7 Re:

are YOU serious?

You realize a beacon packet and an SSID are two different things? And you realize when you claim to be knowledgeable on a subject and make condescending comments about how people must be gobbling up Google PR that posting inaccurate technical descriptions makes you look like a tool ? =)

eric says:

Re: Re: Re:8 Re:

I’ll apologize once more for my sloppy use of language, and my confusion of beacon packets with SSID broadcasts. My mistake.

Once you accept my apology, you’ll see that I actually do have a very good understanding of the technology and what happened.

And you’ll see that my point stands.

Anonymous Coward says:

Re: Re: Re:9 Re:

So, going back to your original point: is there any difference between collecting the unencrypted payload from networks with hidden SSID and collecting the unencrypted payload from networks without hidden SSID? Either both are evil or both are not evil, there should be no difference.

If they were collecting control packets to try to find the SSID for networks with hidden SSID, then there would be a difference (if you are hiding your SSID, it is your misguided intent to have it not known). But what they collected was the content of the data packets, which does not reveal the SSID at all, so the effect should be the same independent of whether the SSID is being broadcast or not.

(Nitpicker’s corner: by SSID I of course mean the ESSID here, not the BSSID.)

Anonymous Coward says:

Re: Re: Re:9 Re:

So, going back to your original point: is there any difference between collecting the unencrypted payload from networks with hidden SSID and collecting the unencrypted payload from networks without hidden SSID? Either both are evil or both are not evil, there should be no difference.

If they were collecting control packets to try to find the SSID for networks with hidden SSID, then there would be a difference (if you are hiding your SSID, it is your misguided intent to have it not known). But what they collected was the content of the data packets, which does not reveal the SSID at all, so the effect should be the same independent of whether the SSID is being broadcast or not.

(Nitpicker’s corner: by SSID I of course mean the ESSID here, not the BSSID.)

eric says:

Re: Re: Re:10 Re:

I don’t know if there’s much of a distinction between recording payloads from networks with hidden SSID and recording payloads from networks without hidden SSID. But you can imagine that a person has a higher expectation of privacy in his traffic on a hidden SSID network than a person on a network broadcasting the SSID. And as such there might be a very important legal distinction.

Google recorded entire packets (including all header bits), so they did actually get SSIDs (and MAC addresses) as well as payload data.

Sneeje (profile) says:

Re: Re: Re:11 Re:

If they have a higher expectation of privacy, then they are ignorant and don’t understand even the fundamental concept of wireless networks.

There is no difference between me running around a public place shouting my bank account numbers with a nametag on, and me doing the samething without a nametag on. I’m still spewing my information into public access spaces.

eric says:

Re: Re: Re:12 Re:

On a normative level, I tend to agree. I don’t think people are careful enough about protecting their own internet communications, nor do they care much about how it works (though I think they should). But whether or not you think what Google did was OK… the legal question is totally different. Based on the information currently publicly available, I believe Google broke the law. Obviously some people disagree with me. But it’s a close enough question to warrant investigations by 18+ countries and 30+ states, plus Congress, the FTC, and the FCC.

Anonymous Coward says:

Re: Re: Re:13 Re:

What we are saying is that it does not matter whether the SSID was hidden or not in relation to the question of whether Google broke the law or not. It is a useless red herring.

Not to mention that when capturing the data packets you do not necessarily know whether the SSID is hidden or not (it is on the beacon packets, a completely different class of packets, and you can easily be too far from the AP to capture the beacons but near enough the STA to capture the data) – one more reason why it should be considered irrelevant to the real question.

Eric says:

Re: Re: Re:

That depends on your definition of “sensitive.” For example, all facebook chats are sent in the clear. Even if you yourself don’t send “sensitive” messages via facebook chat, you would probably admit there are tons of people who do.

Also, putting the onus on a user misses the point. The point is that Google likely violated the Wiretap Act. Regardless of whether or not their activity was normatively harmless, if a court finds that the traffic Google captured is not “readily accessible to the general public,” what they did was illegal.

Anonymous Coward says:

Re: Re: Re:2 Re:

Turning encryption on in your router only encrypts data from your computer to the router. Once it hits the internet it’s in the clear unless you have an encrypted connection to the webserver you are accessing (think https)). Someone could sit in the middle and read chats from any computer to facebook, though this is very illegal as there is an expectation of privacy here.

Expectation of privacy with unencrypted wireless is more nebulous. If you ask me, unencrypted wireless is like standing on your rooftop shouting to your neighbors. You can’t get too P.O.’d if someone walking by heard what you said.

Anonymous Coward says:

Re: Re: Re:4 Re:

There is little expectation of privacy for any data on facebook, but I’m talking about intercepting the message on the way from my computer to their server. I know that the instant my proverbial Aunt Mildred is told something the whole world will know, but if I send her a letter, the post office had damn well better not open it.

eric says:

Re: Re: Re:2 Re:

Ahem, “putting the onus on a user misses the point. The point is that Google likely violated the Wiretap Act. Regardless of whether or not their activity was normatively harmless, if a court finds that the traffic Google captured is not “readily accessible to the general public,” what they did was illegal.”

eric says:

Re: Re: Re:4 Re:

Notice the conditional clause:

if a court finds that the traffic Google captured is not “readily accessible to the general public,” what they did was illegal.

I agree there is disagreement on what is “readily accessible to the general public,” and that the definition of such will probably determine the case. But I never claimed that the payload data was, beyond a shadow of a doubt, not “readily accessible.” To say that I did say that is disingenuous.

Anonymous Coward says:

Re: Re: Re:

or the more likely scenario of in the development process the following logic took place:

In order to get the most accurate geo-positioning information we MIGHT need more than the 24 bytes of the packet so I should write that data out.

OK

Hmm IF packet is encrypted than there is nothing we can use for improving the accurateness of our Geo-location so lets toss that out.

OK

Works great! ship it!

Without the followup that would show-
We have no use for anything beyond the first 24 bytes of any packet so that’s all we need to write to disk

eric says:

Re: Re: Re: Re:

I agree, that’s probably what happened. Look, I’m not trying to say Google did something malicious or that they are actually looking through the traffic they captured in order to spy on people.

But even if your story is exactly what happened, Google still may have violated the Wiretap Act. The Wiretap Act requires no more than an “intentional interception” of data not “readily accessible to the general public.” The second point has not been decided, but the “intentional interception” bit is pretty cut and dry once you read the source code review.

JC says:

Re: Re:

You can look at this patent if you want to see one of many reasons that Google may have chosen passive scanning.

If Skyhook was such a wonderful company where everyone is filled with warm fuzzies all day because they respect privacy so much, maybe they wouldn’t have tried to lock-up technology which allows people to collect wifi data using an “active” scanning method.

eric says:

Re: Re: Re:

I have that patent printed out next to me, highlighted. I understand why Google would use passive scanning. But that misses the point. The point is that Google may have violated the Wiretap Act.

And your ad hominem Skyhook attack is totally irrelevant. We’re talking about the differences in architecture for their data gathering software, nothing more. In that context, Skyhook does a MUCH better job of respecting users’ privacy.

Anonymous Coward says:

Re: Re:

First of all, your link is to an outdated speculative technical analysis.

You’re right that the article doesn’t seem to be written by someone with access to the source code or any data logs, but the author does seem to know what he is talking about. Also, his piece seems to be more about explaining what Google was doing (mapping networks to provide a pseudo-gps capability), roughly how they were they doing and the potential pitfalls rather than the gritty details of how the specific program worked.

You can see that in the source code for gslite (Google’s collection software), it specifically discards encrypted data while writing unencrypted data to disk. It is not reasonable to argue that Google “accidentally” coded their software this way.

Looking more closely you can see that by default the program is written as you say, but the only difference between whether encrypted and unencrypted data is recorded is the setting of a flag. It’s a 3second change one way or the other. Additionally, the default behavior can be overwritten by command line arguments when the program is started.

if you look at Google’s original blog post you can see that gslite was written for a completely different project, perhaps one where recording data packets was useful and fully legal, ethical, etc. It was then appropriated for street view a year or so later (no reason to reinvent the wheel). What should have happened is that either (a) the source code could be changed to set the “discard_data_frame” flag to “false” just like the “discard_encrypted_body” flag or (b) the program that calls gslite (I doubt they’re running it manually) should have had the command line argument to discard all data past the packet header.

Obviously this was not done and someone is in a lot of trouble as they well should be. Not setting a command line argument, however, is a much easier mistake to make than intentionally writing a piece of software to only work one way as you imply.

eric says:

Re: Re: Re:

I didn’t mean to imply that gslite only works one way. But even if it works 700,000 different ways, the only way that matters is the way it worked in practice. And in practice, gslite stored data that was transmitted in the clear.

You’re right, configuring a command line argument incorrectly is an easy mistake to make. But Google was doing this for over three years. Why didn’t they correct it when the data they were capturing included a whole bunch of extra payload data?

One more thing, if you’re relying on Google to tell you the whole story… then you’re not getting the whole story.

eric says:

Re: Re: Re:2 Re:

You’re right that proving knowledge is difficult. But right now the only “evidence” that they didn’t know is what Google says. And there is plenty of circumstantial evidence suggesting they did. (For example, gslite was coded to collect payload data by default; Google didn’t configure the software to discard the payload data; the street view cars were out in the field for over three years; the code went through code review; the size of the database was surely a fair bit larger because of the capture of payload data; etc.) Now, compare that list of evidence to “Google says it was an accident” and “well they didn’t NEED the data, so it MUST have been a mistake.”

Ben Robinson says:

Re: Software development

I am a software developer and as soon as i read about this i could imediately understand how this could happen by accident as Goodle claim. Most of the people claiming this must have been done on purpose clearly have no idea how software is developed. Their main argument seems to be that google wouldn’t write code to collect and store this extra data if they didn’t intend to, but this seems to be based on the idea that every line of code in every piece of software is specific to that software and this is simply not how software is developed. Every piece of software is developed in a modular fashion and you always create a library or collection of libraries that actually do all the nitty gritty work. this is all wired together to make the application. The first thing you do when writing your libraries is think is there already a library that does what i want to do, if there is then you use it rather than reinventing thw wheel. This might be a third party library or it might be something you or your coleagues wrote for a previous project. Every line of code may be reviewed, but only in newly written code, you don’t review every line of mature libraries already in production in other projects. If you did nothing would ever get done. It would be easy to include a library that collected the data you needed, without fully exploring if it might collect extra data that you didn’t need but was required in the original project it was designed for. Software QC is generally looking for bugs and trying to assess whether the software does what it should in a stable and reliable manner. This kind of thing can easily get overlooked as it is actually more of a political issue than a technical one. Us geeky types tend to be highly focussed on solving the problem at hand and can things like collecting a few unneeded random data payloads as trivial and irrelevent.

MD (profile) says:

Unencrypted data transmissions beg to be looked at. If you fail to encrypt your data transmissions then you’re asking for problems. I suppose there is one big issue here:
– what degree of privacy (if any) do you have by allowing your data to be transmitted over an open connection

Considering how up-to-date the 4th Amendment is, I’m sure this isn’t going to be a problem… Not.

So, its unclear whether or not Google intended on capturing the data or not. That is irrelevent. What people are going to have an issue with, is the fact that data WAS captured and they don’t know what will happen to the data – anything from destruction to misuse by a disgruntled employee are probable. The latter is what runs through the minds of people when they realize data they were sending was captured.

As far as privacy is concerned, this doesn’t seem to be so much a privacy issue as perhaps a wiretap issue or even a trespassing issue – altho I’m sure open Wifi could be considered public so it may not even be an issue in that case.

Bottom line is people expect something as “invisible” as wireless data transmission to actually be private. Why would anyone go to Panera or Starbucks to perform credit card transactions in the first place?

Anonymous Coward says:

i would have to say once again you are sounding like a google apologist. i find it amazing you keep posting such articles without pointing out your involvement with google, that you have close personal relationship with key people at google, and that they have allowed you to use their facilities for techdirt business.

based on all of that, your post is suspect. this is the same standard you would apply to others, i am just applying it to your own posts and actions.

JC says:

Re: Re:

stop blathering. this is all i hear when i read your posts.

“blah blah. vague reference to an illicit relationship. blah blah. there must be more to the story.”

Techdirt has repeatedly disclosed their relationship with Google, there are like 10 links on the front page alone to posts where they talk about the work they’ve done with Google.

Anonymous Coward says:

Re: Re: Re:

“vague reference to an illicit relationship” – there is no indication of an illicit relationship. there is indication of a relationship, one which could sway mike on how he reports about (his friends at) google. it is why main stream media is always quick to reveal relationships when talking about companies. you can consult the cnbc website for good examples of disclosure (ge is often in the news, but is also their parent company).

“Techdirt has repeatedly disclosed their relationship with Google, there are like 10 links on the front page alone to posts where they talk about the work they’ve done with Google.” – enlighten me, i dont see any specific links on the front page that is “disclosure of relationship with google” or for any other company for that matter. the only disclosures mike has ever made is after comments like mine, otherwise he ignores the issue and considers it unimportant.

eric says:

Re: Too bad for all the Google haters...

I assume you’re talking about United States v. Ahrndt. Here’s the case: http://scholar.google.com/scholar_case?case=14261433240634475589&q=United+States+v.+Ahrndt&hl=en&as_sdt=20002

While your analysis is close, any decent lawyer can poke holes in it. First of all, it’s a district court case involving child porn. Courts are REALLY reluctant to let pedos off because of Fourth Amendment stuff. Second, Google CANNOT violate the Fourth Amendment (since it only applies to the gov’t). Google’s liability would be under the Wiretap Act, which uses “readily accessible to the general public” language, rather than “reasonable expectation of privacy” language.

While I agree that the case would support Google’s position, it’s by no means a slam dunk.

Dark Helmet (profile) says:

Re: Re: As long as we're talking conspiracy theories...

Yes, see the section “Corporate Members” in the wiki link below:

http://en.wikipedia.org/wiki/Council_on_Foreign_Relations#Corporate_Members

Not that all that much should be made of such membership on the corporate level. In reality, pretty much every major multinational corporation is on there. It’s part of being a MNC at this point, and the corporate members aren’t thought to have much influence within the group, unlike select high ranking members that always seem to turn up in government.

Still, the fact that they’re networking with these people is less than pleasant….

darryl says:

Still an accident - or gross incompetence by a leader in the field of data.

yes, this is the google that is supposed to be upcoming experts in software development, with their own operating system, and application suite. As well they are supposed to be the experts in **Responsible** data mining, and collection, and distribution.

And you are saying that even when they are in trouble for sniffing open wifi in one country, and admiting they did it.

They were engaged in exactly the same practice in other countries. And you (and some blogger, refered to by slashdot), who explains who a packet sniffer works, and you think google would not know exactly what they were doing, and be able to only read the 24 bytes of the packer, and the signal strength info.

You say they do not use the data, then why have I seen google maps with open and closed WiFi ports overlayed ?

To say a company like google “accendently” collected that data is just an attempt at mitigating their actions, and trying to take away to premeditation that is required to undertake this act.

The other thing is “there is no such thing as an accident”.

“An accident is a specific, unidentifiable, unexpected, unusual and unintended external action which occurs in a particular time and place, with no apparent and deliberate cause but with marked effects.”

You just cannot have multiple exactly the same “accidents” in different countries, by different people, employed by the one company.
That is not an accident, its premeditated, deliberate, and clearly done for a specific purpose and outcome.

Considering it is so easy to correctly gather the **right** data, and not gather this information.
And a company like Google, are too big, and too expert in this field to claim it as an accident.

Sure, I concede that it was an accident for google to do it in the first place, and to get caught second.

Its a major accident to googles reputation as a company that acts responsibility and that is aware of the importance of privacy, and the responsible dealing with vast quantities of data. (that is not theirs).

So if a Mike, or a Joe blogs claims to accendently log open WiFi, I would not believe him either, but it is more feasible someone could do that one time by mistake, but not a Google, who while in trouble in one country for this perform the exact same act in several other countries.

That is not an accident, and Google is the last company I want having “accidents” with my data, thankyou.

So the blog, that slashdot quoted, that you quoted from slashdot, (thats Robert Graham of actual content).
He is a blogger like mike, what he says, makes no bearing on Googles acts. But Robert does show how it is trivial to do it right, which would imply a company like google would be fully aware of that as well.

No accident there.

and no explination of how it could possibly be an accident..

An accident is a unique event (at a particular place and time), it is not possible to have ‘an accident’ in multiple different countries at different times.

They are not the same accident, each one is an independent accident, which is not feasible.

So again, Googles claim it was an accident, is like asking a thief if he is guilty or not. Ofcourse, their opinion is highly biased.

Its up to observers to determine accident, or premeditation, and clearly repeating the same ‘accident’ after being called out for it in one country, must not be an accident.

An accident is something that happens that was not your INTENT.
You can drive a car, your intent is to get to your destination, an accident is something happening, unexpected and not predictable.

If you are driving far beyond the speed limit, you are not expecting to have a crash, but if you do crash it was not accendental, as you were driving wrongly. (too fast).

you can never perform that same event, that accident (caused by intent, by speeding), as an accident is something that occures at a particular place and time.

Are you saying that the google people spent months driving around all these sites, gathering data, and during that time did not review their work to date, and notice this information.

Ofcourse they did, and ofcourse they were away that data was provided, I assume under managements instructions.

You do not map vast areas of a country, and logs huge amounts of data and not test, review and examine the data and collection method until the end of the project.

You test these at the begining, prove the system before using it.

Your saying a company the size of google would do all this collection, and not MEAN to collect, and not notice DURING the collection that certain data was being collected ?

I guess claiming it was an accident might please some, but for most its beyond the possibility that google did not know what they were doing it, while they were doing it, and the multiple countries they were doing it in !!!..

but if you believe that, that is fine..

Go ahead trust google with everything you own.. for me,, no thanks.

Saying it was an accident means they are either lying or incompetent, take your pick ..

Mike Masnick (profile) says:

Re: Still an accident - or gross incompetence by a leader in the field of data.

You say they do not use the data, then why have I seen google maps with open and closed WiFi ports overlayed ?

I answered this last night in response to your other comment on this topic. They were mapping WiFi networks. That is not illegal. What you saw was perfectly legal and many companies do it.

The questions here are not about mapping WiFi. It’s about the collection of data going through those routers. No one denies that they were mapping WiFi data points. But that’s legal.

Consider we explained this to you last night, I cannot figure out why you would continue to repeat the same falsehood.

You just cannot have multiple exactly the same “accidents” in different countries, by different people, employed by the one company.

Again, this was explained to you last night. Of course you can: all Google locations were using the same technology, so of course the same thing happened everywhere.

Sure, I concede that it was an accident for google to do it in the first place, and to get caught second.

They didn’t “get caught.” They admitted it publicly once they found out about it.

An accident is a unique event (at a particular place and time), it is not possible to have ‘an accident’ in multiple different countries at different times.

Darryl, can you please stop and listen for one second: the *accident* was in setting the software to record the unencrypted packets. That happened once. But that was in the software that was rolled out to all the street view vehicles. So, yes, the “accident” happened once, but there were repeated uses of it afterwards, without people realizing it.

Let me ask a question: with all the reports of Toyotas automatically speeding up that were in the news a few months back, leading to widespread recalls, based on your reasoning above, this must have been “intentional” by Toyota, right? After all, it happened many times around the globe. By your reasoning here, Toyota “intentionally” planned to make its cars accelerate without warning.

But, of course, that’s ludicrous. It seems clear that there is some flaw in Toyota’s setup. That flaw is an accident. But it was rolled out to cars worldwide.

Same thing with Google.

Are you saying that the google people spent months driving around all these sites, gathering data, and during that time did not review their work to date, and notice this information.

They did review the info they were TRYING to collect. The location of WiFi hotspots. That’s not the same as reviewing the payload data.

You have, in the past, claimed that you are a well known technologist. I’m having trouble believing that when you seem to be confusing some very basic technology issues.

darryl says:

Re: Re: Still an accident - or gross incompetence by a leader in the field of data.

You are appoligising for google, as if they suddenly decided to one afternoon go out and collect these WiFi points.

Then after they did that, and got home they looked at the data and said “oh dear, we logged the entire country, taken weeks or months to do it, and only now that we look at the data we see we got too much”.

Do you **NOT** think Mike they might do some small trials first may be a small city, then look at the data, and work out how to put it on a map ?

Or;

Do you think they went out and did the entire country, just **ONE** group, in a short period of time, and during that time did not look at their data to date, even just to check the system was working.

AND;

Do you expect us to believe that someone who is supposed to be technology savvy would make such a huge error, or undertake such a huge project without testing and trials first ?

AND finally; do you expect us to believe that ?

No, im not a well known technologist, im a knowlegable technical engineer.

and unlike the fellow you qoute here as your souce, I do not claim that “im a world leader in this (WiFi) technology and have been studying it for the past 20 years”

This guy you quote, has been studying WiFi technology since before it existed !!! amazing.

A car company that produces a part that is faulty, is an error not an accident. By definition.
Again, you dont have multiple accidents.

That is just wrong, as you should well know. You use words all the time Mike, you should use the right ones, and be accurate on using the right words to describe the right events.

I know you like to spin to a degree, and like to use violation over theft, (tell a 6 year old who has been violated why that is better than theft ?)

A single component, that is produced, that has an impurity or something falls into the machine is an accident.

A design error, that results in many faults is not an accident. Its not unexpected, if the designer had of done the right job in the first place the error would not occur.

In an accident, the designer can do everything right, but the part can still fail. Can you or anyone else see the difference.

You design a bridge wrong, and it falls down its not an accident, its criminal neglegence (or gross incompetence).
The bridge did not fall down accendently, it fell down because it was designed or built wrong.

So you cannot have multiple accidents, you can have multiple errors.

And you Mike, who tries to be a word smith, should take more care and accuracy in your statements. Should be desire to be taken seriously.

Mike Masnick (profile) says:

Re: Google apologia?

Nothing like waking up in the morning to find a near totally fact free hatchet job from Scott Cleland.

First, I find it amusing that he still hasn’t even noticed or responded to my response to his first totally false hatchet job. But that’s typical. His job is not to respond to facts, but to lie about people who are actually getting it right.

As for what he said? Let’s take a look.

Last week Mr. Masnick summarily dismissed the possibility that Google could behave anti-competitively in search

I did not such thing and already responded to this assertion. Saying that there is no such thing as search neutrality is NOT the same thing as saying Google couldn’t behave anti-competitively.

Why would Cleland blatantly lie? I don’t know.

Apparently Mr. Masnick has drunk the Google Kool-Aid since he is sure that Google could not do wrong in the antitrust and privacy arenas.

Again, I neither said nor implied any such thing. In fact, I have dinged the company on privacy issues in the past multiple times. I believe the company very much can and has done wrong on those things. But that doesn’t mean I believe every fear mongering story that AT&T paid Cleland to write.

Mr. Masnick would retain more credibility if he acknowledged the possibility of Google’s fallibility in these human endeavors.

I have and continue to do so. Amusingly, in Cleland’s rush to say I never argue against Google, he appears to miss the two posts last week where I did exactly that. For example he totally missed this post: http://www.techdirt.com/blog/wireless/articles/20100610/1358039772.shtml

In last week’s blind defense of Google that “There Is No Such Thing As Search Neutrality, Because The Whole Point Of Search Is To Recommend What’s Best,” Mr. Masnick effectively dismisses the possibility that Google could have monopoly power

I did no such thing. I don’t know why he continues to say this. I also already responded to this charge in a post. Cleland repeating it is just kinda sad.

even though the U.S. DOJ Antitrust Division determined Google is dominant in search advertising and was prepared to bring a Sherman Act monopolization case against Google in November 2008, if it did not drop the proposed Google-Yahoo ad Agreement.

That has nothing to do with what I actually wrote about: which is whether or not “search neutrality” exists.

Mr. Masnick apparently does not understand the concept in antitrust that a company with market dominance or monopoly market power must operate under a different and stricter standard of conduct to prevent anti-competitive behavior.

I do actually understand that quite well. And I, unlike Scott Cleland, also seem to understand that this has absolutely nothing to do with what I actually wrote about.

But, since Cleland apparently is unwilling to deal with what I actually said, he needs to blatantly lie about what I said.

Note, by the way, that Cleland doesn’t actually quote anything I said. He just pretends I said all sorts of stuff I never said.

Last week I rebutted Mr. Masnick’s Google antitrust apologia by explaining that even Google itself disagreed with Mr. Masnick’s characterization

No, as I already pointed out (http://www.techdirt.com/articles/20100621/0355239887.shtml — though Cleland apparently didn’t even bother to read it). What Cleland did was pretend I said something I did not. He pretended “search neutrality” meant something different than it does and pretended I made claims that I did not. When I responded, Cleland did not update or change his position. He just keeps repeating it.

Keep it up Cleland. It just makes you look that much more ridiculous.

I challenged Mr. Masnick because he was completely ignoring a serious and legitimate issue, search neutrality, and also ignoring substantial evidence in the public domain of a Google antitrust problem.

And I responded, but apparently Scott Cleland doesn’t actually read responses to his lies.

Now this week Mr. Masnick is assuming the role of Google’s go to apologist again by declaring he is “almost certain” Google’s three-year WiFi data collection in over thirty countries by retrofitting its entire StreetView vehicle fleet with special WiFi antennae was “accidental.”

Now this is the fun part. Cleland doesn’t actually read or respond to what I wrote. He doesn’t note that I linked to and quoted a security expert who explained in great detail what likely happened.

No…. instead, he seems to base his whole complaint on the headline, and makes it sound like I’m just giving my opinion without any factual basis to back it up at all.

Apparently Cleland assumes that his readers are so braindead that they don’t actually read source material.

Has it ever occurred to Mr. Masnick that Google’s well-known mission to organize all the world’s information at least suggests the possibility that this particular WiFi data collection could have been intentional and not an accident, given that Google tracks most everything else that happens over the Internet?

Sure, that’s occurred to me. But we weren’t randomly speculating like you do. We looked at the actual evidence. Something Cleland is not fond of doing.

Mr. Masnick would have more credibility if he did not summarily dismiss the possibility that someone in the privacy offices of several nations, the FTC, Scotland Yard, and roughly 30 state Attorneys General who are now investigating Google’s international WiFi data collection, might know something that Mr. Masnick might not be aware of?

Unlike Scott Cleland, I’m willing to wait until after these things go through before assuming what these investigators know or do not know.

I find this particular Clelandism telling. Multiple times now, he attacks me for ignoring evidence, but each time that “evidence” is merely a charge by either a gov’t official or another company that Google *MIGHT* be doing something. To Cleland, just that is evidence. That Google has yet to actually face any of these charges in court apparently is lost on Cleland. Odd.

In short, Mr. Masnick would enjoy more credibility on these topics if he spent less time parroting the absolutist Google PR talking points and more time being “open” to the mounting evidence that Google may be acting anti-competitively and may be acting irresponsibly concerning consumers’ privacy.

Except, of course, my points have nothing to do with Google PR.

Scott, just because you are well-paid to spout anti-Google talking points, you really shouldn’t assume that anyone who proves you wrong is doing the same thing.

And, in the future, when you choose to write an entire hatchet job post attacking me, you could at least respond to what I actually wrote, other than what you pretended I wrote.

This, by the way, is why no one in DC even pays attention to Cleland any more. He’s done this so many times his name is a joke.

rl78 says:

General question regarding this issue....

I don’t understand wireless data transmissions and software development thoroughly enough to formulate my own opinion whether capturing these payloads packets was intentional or not but I still have a question maybe someone can answer.
From what I understand or from what this article states, Google only needed MAC addresses and signal strength data from these unsecured wifi networks. This was sufficient for what it wanted to do, but because they were using some code originally designed for another project, in this instance it ended up capturing more data than it needed, and stored it.
Google continued to do this for 3 years. Its the length of time they were doing it which gets me because as soon as it was “brought to their attention”, they apologized, said it was accidental and not only agreed to stop sniffing payload packets, they decided to stop sniffing altogether.

Let’s say it was accidental because of the code borrowing (if I can say it like that), and they didn’t re-review the already functioning code. Wouldn’t they have reviewed the data they captured and noticed the unwanted, or unneeded data long before it was brought to their attention by another party?

If yes, then weren’t they just leaving themselves open to being looked at as invading privacy by not modifying the code to not do this? If they were doing nothing illegal by simply storing that data, then shouldn’t they just stand behind their practice? I could see the argument to stop it all just because its not worth the bad pr, and users may still perceive it as an invasion of privacy and move away from Google, but by apologizing for it, and then stopping altogether, seems to me like an admission of wrongdoing of somekind in itself, making the apology warranted, but then leaving the question was it intentional or not.

If no, and there are people out there who think or know that Google could have done this for 3 years without realizing it was capturing data that might either get them in trouble or was legal but would still cause them a headache, then alright. I really don’t know enough about it to say otherwise. This would also most likely be their defense if it was illegal.

In my limited opinion, its hard for me to believe they simply never noticed they were collecting this data during the 3 years, but this is why I ask?

Mike Masnick (profile) says:

Re: Bias Detector

Couldn’t help but notice “Ads by Google” banners and doubleclick.net (Google subsidiary) ads bordering your post.

Do you have something substantive to add? I have regularly sided against Google on this site. The idea that the few dollars I get from Google each year (significantly less than 1% of our revenue) influences the coverage is preposterous.

Necroboy (profile) says:

Quite a lot of the above

Hey, I came across this old thread after today’s ruling on Google’s collection of unencrypted data data via their Streetview project and just couldn’t resist giving it belated CPR.

I was wondering what those commenting above might have to say in light of recent revelations regarding Google’s (and others’) alleged role in enabling wholesale incursions into citizens’ private lives by our security organisations?

Leave a Reply to TtfnJohn Cancel reply

Your email address will not be published. Required fields are marked *

Have a Techdirt Account? Sign in now. Want one? Register here

Comment Options:

Make this the or (get credits or sign in to see balance) what's this?

What's this?

Techdirt community members with Techdirt Credits can spotlight a comment as either the "First Word" or "Last Word" on a particular comment thread. Credits can be purchased at the Techdirt Insider Shop »

Follow Techdirt

Techdirt Daily Newsletter

Ctrl-Alt-Speech

A weekly news podcast from
Mike Masnick & Ben Whitelaw

Subscribe now to Ctrl-Alt-Speech »
Techdirt Deals
Techdirt Insider Discord
The latest chatter on the Techdirt Insider Discord channel...
Loading...