Archive for November, 2009

Lecture David Gugerli – The Culture of the Search Society

Posted: November 30, 2009 at 5:28 pm  |  By: margreet  |  Tags: ,

Data Management as a signifying practice
David Gugerli, ETH Zurich
November 13, 2009, Amsterdam

Edited by: Baruch Gottlieb

Databases are operationally essential to the search society. Since the 1960’s, they have been developed, installed, and maintained by software engineers in view of a particular future user, and they have been applied and adapted by different user communities for the production of their own futures. Database systems, which, since their inception, offer powerful means for shaping and managing society, have since developed into the primary resource for search-centered signifying practice. The paper will present insights into the genesis of a society which depends on the possibility to search, find, (re-)arrange and (re-)interpret of vast amounts of data.

Download here the full lecture of David Gugerli given during the Society of the Query conference on Friday the 13th of November 2009.

Siva Vaidhyanathan on Googlization, “Only the elite and proficient get to opt out”

Posted: November 19, 2009 at 7:13 am  |  By: admin  |  Tags: , ,

Society of the QueryThe term Googlization, according to Siva Vaidhyanathan, is the process of being processed, rendered, and represented by Google.

Vaidhyanathan’s upcoming book The Googlization of Everything investigates the actions and intentions behind the Google corporation. This afternoon at The Society of the Query Vaidhyanathan choose one issue from his book: the politics and implications of Google Maps’ Street View.

According to EU law: there cannot be any identifiable information about a person in Google Street View. Google’ s standard defense up till now has been that they respect privacy by scrambling faces and license plates, to which Vaidhyanathan commented,

In my former neighborhood in New York there were probably 50 illegal gambling institutions around. Now, imagine an image of me on Google Street View taken in proximity to one of these illegal places. I’m more than two meters tall and I’m a very heavy man. You could blur my face forever, I’m still bald. In New York, usually I was walking around my neighborhood with a white dog with brown spots, everyone in the neighborhood knew that dog. So you could blur my face and it still wouldn’t matter – it’s me, I’m obviously me. Anonymization isn’t an effective measure, as we’ve already found out with data. (most likely referring to the AOL case of user #4417749)

Just this morning Swiss authorities made a statement that they plan on bringing a lawsuit against Google in the Federal Administrative Tribunal because Google isn’t meeting the country’s demands for tighter privacy protection with Google Street View. Vaidhyanathan commenting on the news said, “Google Street View has been entering so many areas of friction and resistance – this brings it to our attention that the game is over for Google.”

Vaidhyanathan’s criticism of Google Street View continued with Google’s trouble in Tokyo. “The strongest reaction against Google Street View has been in Japan,” he said, “Google will scrap all of their data from Japan and re-shoot the entire country. Google mismeasured how the Japanese deal with public space. In the older sections of Tokyo the street in front of one’s house is considered the person’s responsibility, it is seen as an extension of their house. Thus, Google Street View is actually invading someone’s private space.”

Earlier this year Google CEO Eric Schmidt made the following remark about the international appeal of Google,

The most common question I get about Google is ‘how is it different everywhere else?’ and I am sorry to tell you that it’s not. People still care about Britney Spears in these other countries. It’s really very disturbing.

Vaidhyanathan explained this as being a part of Google’s protocol imperialism,

Google isn’t particularly American, nor is it particularly American / Western European. It’s important to remember that Google is much more a factor of daily life in Europe. In the United States it is just barely 70% of the search market, in Western Europe it is around 90% and in places like Portugal it is 96% and I don’t know why.

For Vaidhyanathan the biggest problem with Google is that as it expands into more parts of the world that are less proficient, and less digitally inclined, there will be more examples of friction and harm because more people are going to lack the awareness to cleanse their record.

It’s important to note that Google does offer services for protecting and managing user data:

Vaidhyanathan didn’t specifically mention these options, but briefly acknowledged the existence of such tools before quickly moving onto the strongest part of his argument, “We in this room are not likely to be harmed by Google because all of us in this room are part of a techno-cosmopolitan elite. Only the elite and proficient get to opt out.”

Google Street View Fail

In closing, Vaidhyanathan exemplified the problem with a photograph of a man caught on the side of a U.S. highway and commented, “This man doesn’t know that he is in Google Street View so we get to laugh at him. Not knowing is going to be the key to being a victim in this system.”

More information about Siva Vaidhyanathan and his criticism of Google can be found on his website, and in this lively Google debate at IQ2 and New York Times article from last year.

Siva Vaidhyanathan on Googlization, "Only the elite and proficient get to opt out"

Posted: November 19, 2009 at 7:13 am  |  By: chris castiglione  |  Tags: , ,

Society of the QueryThe term Googlization, according to Siva Vaidhyanathan, is the process of being processed, rendered, and represented by Google.

Vaidhyanathan’s upcoming book The Googlization of Everything investigates the actions and intentions behind the Google corporation. This afternoon at The Society of the Query Vaidhyanathan choose one issue from his book: the politics and implications of Google Maps’ Street View.

According to EU law: there cannot be any identifiable information about a person in Google Street View. Google’ s standard defense up till now has been that they respect privacy by scrambling faces and license plates, to which Vaidhyanathan commented,

In my former neighborhood in New York there were probably 50 illegal gambling institutions around. Now, imagine an image of me on Google Street View taken in proximity to one of these illegal places. I’m more than two meters tall and I’m a very heavy man. You could blur my face forever, I’m still bald. In New York, usually I was walking around my neighborhood with a white dog with brown spots, everyone in the neighborhood knew that dog. So you could blur my face and it still wouldn’t matter – it’s me, I’m obviously me. Anonymization isn’t an effective measure, as we’ve already found out with data. (most likely referring to the AOL case of user #4417749)

Just this morning Swiss authorities made a statement that they plan on bringing a lawsuit against Google in the Federal Administrative Tribunal because Google isn’t meeting the country’s demands for tighter privacy protection with Google Street View. Vaidhyanathan commenting on the news said, “Google Street View has been entering so many areas of friction and resistance – this brings it to our attention that the game is over for Google.”

Vaidhyanathan’s criticism of Google Street View continued with Google’s trouble in Tokyo. “The strongest reaction against Google Street View has been in Japan,” he said, “Google will scrap all of their data from Japan and re-shoot the entire country. Google mismeasured how the Japanese deal with public space. In the older sections of Tokyo the street in front of one’s house is considered the person’s responsibility, it is seen as an extension of their house. Thus, Google Street View is actually invading someone’s private space.”

Earlier this year Google CEO Eric Schmidt made the following remark about the international appeal of Google,

The most common question I get about Google is ‘how is it different everywhere else?’ and I am sorry to tell you that it’s not. People still care about Britney Spears in these other countries. It’s really very disturbing.

Vaidhyanathan explained this as being a part of Google’s protocol imperialism,

Google isn’t particularly American, nor is it particularly American / Western European. It’s important to remember that Google is much more a factor of daily life in Europe. In the United States it is just barely 70% of the search market, in Western Europe it is around 90% and in places like Portugal it is 96% and I don’t know why.

For Vaidhyanathan the biggest problem with Google is that as it expands into more parts of the world that are less proficient, and less digitally inclined, there will be more examples of friction and harm because more people are going to lack the awareness to cleanse their record.

It’s important to note that Google does offer services for protecting and managing user data:

Vaidhyanathan didn’t specifically mention these options, but briefly acknowledged the existence of such tools before quickly moving onto the strongest part of his argument, “We in this room are not likely to be harmed by Google because all of us in this room are part of a techno-cosmopolitan elite. Only the elite and proficient get to opt out.”

Google Street View Fail

In closing, Vaidhyanathan exemplified the problem with a photograph of a man caught on the side of a U.S. highway and commented, “This man doesn’t know that he is in Google Street View so we get to laugh at him. Not knowing is going to be the key to being a victim in this system.”

More information about Siva Vaidhyanathan and his criticism of Google can be found on his website, and in this lively Google debate at IQ2 and New York Times article from last year.

Martin Feuz (CH) Google Personal Search – What are we to make of it?

Posted: November 17, 2009 at 10:57 pm  |  By: rosa menkman  | 

Martin Feuz is an independent researcher with a strong interest in human-information interactions. Specifically, he focuses on exploratory (Web) search and the ways in which such interactions can be meaningfully and experimentally supported. In his recent work, he undertook a critical analysis of Google personal search to open the lid of Google’s black box a little bit and to make some of its behavior more door reflection.
In Society of the Query, Feuz presents the research that lead to the development of his new website Perspectoma.com. Perspectoma is a research engine that allows us to get a glimpse into how Google Personal Search delivers ‘personalised’ search results on the basis of an users Search and Web History.
Perspectoma works by simulating Search and Web Histories. The website offers five profiles for which Search and Web Histories are developed. These profiles contain approximately the equal amount of search queries that an average user accumulates over two years of searching with Google.
When performing a search, Perspectoma’s search result page shows:

•    the search results only available for the profile
•    the search results of an anonymous user without personalization
•    the search results that are available to both, the selected profile and the anonymous user and but have a different search result ranking (blue)
•    the search results that are available to both, the selected profile and the anonymous user and share the same search result ranking (green)

Google describes personal search as ‘based on your search & web history. Your personal results will be re-ranked or swapped for more relevant ones.” However, it gives no indications whatsoever when a particular search result is personalized. Therefore you actually never really know where your returns come from, and which ones are specially there to target you as a consumer, or to help you. Google states that if you don’t want personal search, you can just sign out of your Google account, Unfortunately, this is not a very practical because in the end you seem to sign into Google very often and easily forget to sign out.

Feuz starts his presentation by posing four main questions he wanted to deal with while creating Perspectoma.

•    how soon (after how many queries) will Google Personal search play a roll in the Google search returns?
•    how do these personal returns develop in terms of frequency and intensity?
•    what underlying patterns can we identify?
•    how will grouping of profiles influence the search terms?

To find answers to these questions, Feuz describes the research he did according to the 3 ghost accounts based on the characters Kant, Foucault and Nietzsche. He trained all of the accounts equally in training sessions with the help of software that did different search queries. To do this, he had to find a way to make an artificial profile relational for subjective inspection of personalized search results. To tackle this problem, he used specific books for the different theorists. He also had to find a way to get plenty search terms to create profiles.
After training session 1 Feuz found that the search returns for Foucaults profile were personalized results quite early, but not very frequently. The search returns for Kant were a bit of personalized but not to much/to often. Feuz also considers that this could have to do with the type language Kant uses in his books. For Nietszche a lot of personalized results turned up, but this actually was the result of a small glitch in the technology.
Martin Feuz concludes that he is surprised how soon the personal search returns seem start turning up. Google personalized search is not shy. After the second training sessions the amount of personal returns seem to grow, while after 3000 search queries more than every second result is personalized. Also, it seems that there is a kind of group profiling happening.
Finally, Feuz states that personalized search does not seem to be able to lift the less dominant voices from deep down the unbiased search returns. Actually, it seems that most often personalization means that only some of the returns from the second page have been swapped into the first ten personal search returns.

Antoine Isaac: Semantic Search for Europeana

Posted: November 17, 2009 at 5:02 pm  |  By: tjerk timan  |  Tags: , ,

Society of the Query

Thanks for the opportunity to talk. I work at the VU and I am talking about the project Europeana. This is the result of a teamwork of the University, I am just presenting it.

Introduction
What is Europeana? It is a portal which that want to interconnect museum archives. Access to digital content. Currently there are 50 providers. and the number is growing. 10 million objects is the target. Maybe from a more practical : we want to create access but we also want create channels to other websites and so on. Such a thing does not go without challenges. The very issue of providing access that is very difficult. They are of an iterational nature. And how to get data besides the pictures? The method is to use metadata. Antoine shows the current portal, which he explains as a “basic search box” (picture needed). If a search query is done, different result are given that are linked to the search (pic, books etc). You can start refining you search by filtering (language, data and so on). This is called semantic search and it allows you to refine your search. To some extend this is not matching the richness of data that is out there in the databases. The idea is to go a step beyond semantic enables search. Some functionalities are explained, such as clustering. Antoine explains that by exploiting semantics, we can exploit relations that are stored in the objects. We can use information that is actually there already in the meta data. Some kind of organized knowledge is already there, we want to exploit it. The proper information properly accessible , that is the goal.

Example
A semantic layer on top the ‘normal’ results is presented. A graph is shown of a semantic web. It needs to become more useful for users, according to Antoine. A common concept that can aggregate for relations. A screen shot is given of the prototype. It is a mini-version of the total project: three museums are currently represented. You can start typing your search. The first difference ( from normal search engine red) is that it will be able to provide you with concepts and locations that could match your string. If you select one of the results , you get a number of new possible links and clusters via criteria. It is important to notice that the results are coming from really specific entities. We can see that the subject “egypt” for example gives a whole set of related objects. It is much more than a returned string.

This idea of having controlled entities can be used in more complex means. Users can  start exploring further knowledge and concepts.  An example is given on the search “egypt’ and the meta results. We are now searching via concept/relations.  This is an example of richer information. I also got clusters like works created by somebody who was in Egypt and so on… The reason for getting this object in the results is that in the metadata links back to the subject (query). There is a kind of person space emergent here.  Via this person, we can find out the place and we end up in Cairo. One very important point is that we benefit from existing models and vocabularies. Via labels on concepts, these concepts can be linked. It is very important because now you can access this information. We continue by determining these links (exact matches and relational matches). The main advantage of metadata is that it is heterogeneous. There a different description models. You cannot really anticipate it. Some form of alignment is required in order for the system to work, because these databases use different vocabularies. A data cloud is presented which represents the different vocabularies in the three different museums. These vocabularies are glued together.

Conclusions
The semantics in our case are getting structure in the data. It is about coupling the data.. It is a flexible architecture. It is about loading data. This makes ingestion for new data easy.  You don’t need to fully merge the workings of all the institutions/ content providers.  It is about connecting structures together. It allows easier access to the different vocabularies. You can start your search and you are provided with different vocabularies. Next, we have to bring in more vocabularies. You can have quality data in this system.  Finally, this vision  of the variable links model is nice, but some semantic matching level problems occur. This is difficult. A link is given: here you can try the portal here

Questions
Rogers: Don’t you need an army if you want to actually make the links and translation between all the records?
Isaac: you are right, we actually implemented something (the three museums vocabularies), we are not experts on library science. Until recently, however, the library scientist did not come out of their institutions. Now, they start to realize they can integrate their knowledge. I believe this is an added value.

Rogers: Is this more than digitizing library systems? Is this indexible by Google?
Isaac: Yes, it should be.
Rogers: is it deep indexible? isn’t this a huge policy question?
Isaacs: This prototype publishes the data. You can see the source of the data.

Pembleton: analogy: Tim Bernes-Lee created a website that can point to all your data. What I see here is the- same move. By linking the concepts, not the data. This provides a richer web.
Rogers: Is this a Europe-island web, then?
Cramer: We already have such a system: it is called RSS.

Audience: A method that I see here is: we need glue to link existing concepts and vocabularies. The other is to generate new vocabularies . To me that seems to be a large debate.
Pembleton: We use the same underlying technology.  I see more added value rather than competition.
Cramer: RDFA is not a vocabulary, it is a language to format the vocabulary (which is a huge difference).

Michael Stevenson presents a Google art expose

Posted: November 16, 2009 at 4:15 pm  |  By: rosa menkman  |  Tags: ,

Society of the QueryMichael Stevenson is a lecturer and PhD candidate at the Department of Media Studies, University of Amsterdam. For the Society of the Query evening program he presented a very interesting selection of artistic and activist projects that were engaged with (the re-attribution of) different elements related to Web search.

Query

The IP-Browser (Govcom.org) for instance played with the linearity of querying the Web. It creates an alternative browsing experience that foregrounds the Web’s machine habitat and returns the user back to the basics of orderly Web browsing. The IP Browser looks up your IP address, and allows you to browse the Websites in your IP neighborhood, one by one in the order in which they are given in the IP address space.

Shmoogle (Tsila Hassine/De Geuzen) also deals with linearity on the Web, specifically the linearity of the search returns of Google. De Geuzen state that the best search returns that Google offers are not necessarily always the ones at the top. Unfortunately this is where the average Google user gets stuck. Shmoogle offers a way to find the search results in a chaotic way, and in doing so it ensures greater democracy.

The Internet Says No (Constant Dullaart) is a animated, fully functioning Google page (Google is placed in a marquee-frame). this work offers a pessimistic way to surf the internet.

The Misspelling-Generator (Linda Hilfling & Erik Borra). Erik Borra presented the work as a result of the fight against internet censorship. When doing a search in the Chinese version of Google on the Tiananmen Square Massacre, Linda Hilfling discovered a temporary loophole out of the Google self-censorship in China. By deliberately spelling Tiananmen incorrectly, she was taken to web-pages where other people had misspelled Tiananmen, and was thereby able to access pictures of demonstrations as well as the legendary image of the student in front of the tank through the sources of incorrect spellings. The Misspelling generator is a tool that can be used for internet activism. By writing variations like ‘tianamen’ and ‘tiananman’ the isolation politics of the Google’s spelling corrector can be subverted while Google’ selfcensorship can be circumvented.

Society of the Query

Images

Z.A.P. (ApFab) is an automatic image generation installation. First you add a word using the ApFab touch-screen, then the ZapMachine will grab an image from the Internet. This image is the most important visual representation of that word, at that time, according to the current Internet authority Google. Finally, the individual images are incorporated into a new context, creating a new tense state of meaning and random relations. With “Zapmachine: Who gave you the right?” AbFab is asking the following questions:

-How much control do we have over the generated collage as artists?
-How much influence do you have on this process.
-How does the collage relate to the initial intention by which the image was uploaded on the Internet by the original author?
-Who is the author of this Zap collage?

Disease Disco (Constant Dullaart) “To every suffering its thumbnail”. Dullaart used the Google image search by color option, to query the word ‘disease’ and changes color ‘rhytmically’. The work is accompanied by the US billboard #1 hit song of the moment that the work was created.

The Global Anxiety Monitor (De Geuzen) uses html-frames to display automated image searches in different languages. Searching in Google for terms such as conflict, terrorism and climate change, this monitor traces the ebb and flow of fear in Arabic, Hebrew, English and Dutch.

Terms & Conditions

Cookie Monster (Andrea Fiore) To capture on-line behavior, thousands of HTTP cookies are sent daily to web browsers to identify users and gather statistical knowledge about tastes and habits. The cookie consensus website hosts a collection of cookies that Andrea Fiore received while surfing through the first 50 entries of the Alexa directory of News sites. In the future it will also host a software that will give the users the capability to create their own cookie collections.

I Love Alaska (Lernert Engelberts & Sander Plug) is a beautifully framed internet movie series that tells the story of a middle aged woman living in Houston, Texas. The viewer follows her AOL search queries over the time span of months. “In the end, when she cheats on her husband with a man she met online, her life seems to crumble around her. She regrets her deceit, admits to her Internet addiction and dreams of a new life in Alaska.”

Society of the Query

http://www.geuzen.org/anxiety/

Ton van het Hof (NL) about flarf poetry

Posted: November 16, 2009 at 3:02 pm  |  By: rosa menkman  |  Tags: , , , ,

Society of the Query

Flarf poetry can be characterized as an avant-garde poetry movement of the late 20th and the early 21st century. In flarf poetry a poet roams the Internet using random word searches, to distill newly created phrases and bizarre constructions that he later shares with the flarf community.

Flarf poetry can be described as a ‘readymade’, collage technique that has connections to the Surrealists in the 20s and William Burroughs cut-up technique from 1959. Flarf itself exists for a decade and has since then evolved by using web poetry generators and chatbots like Jabberwacky.

YouTube Preview ImageTon van het Hof showed an example of flarf by Sharen Mesmer: “A knowing diabetic bitch”

This is my Readymade Flarf poem using Jabberwacky:

What is Flarf? The greatest two dimensional thing in the world. What is Flarf? A Flatland. It’s a satire on this one.

Although my self made poem doesn’t show this so well (I am unfortunately an amateur flarf poet), flarf poems are often as disturbing as they are hilarious, which have made many people question if flarf will can ever be taken serious. Even though this question is still a valid question today, the movement is showing signs to have cleared a spot amongst the ranks of the legitimate art forms, finding its ways to blogs, magazines and conferences.

Discussion session 2: Googlization

Posted: November 16, 2009 at 12:16 am  |  By: tjerk timan  |  Tags: , , , , , ,

With: Siva Vaidhyanathan, Martin Feuz  and Esther Weltevrede

Moderated by Andrew Keen.

Society of the Query

Moderator: Why does no one talk about money?

Vaidhyanathan: Google only loses money. They have an interest to keeping people interacting with the Web. As long as you are interacting with the web, they can track you via cookies and that puts more data in their database. It is a clear but third degree connection for creating revenue. It also has interest in data- and text accumulation. It hopes to create a real text-based search. In terms of Google search; global and local are not really put to for example; Google books. This already biases the search results.

Weltevrede: It also depends on your perspective on Google. For me it is interesting to see how it works. How does it organize and present the information we get.

Vaidhyanathan: nobody is going to Google for the ads.

Audience (at Weltevrede): you were depending on the Google translation results?  Isn’t that tricky?

Weltevrede: indeed,  Google Translate is still in beta version. However, human rights is such an important term that one can assume that it is translated well.

Society of the Query

Audience: how about methods? It is difficult to pose yourself against the machine. All of us here agree that searching sucks and that Google is bad and commercial. So I’d like to have some reflection on methods in order to be critical against searching and how they relate to your research?

Vaidhyanathan: Google is hard to study in traditional way. I do my best to keep to fuzzy, flabby arguments of narrative and argument. Opacity is the problem of Google. You cannot research is without using it. You risk becoming a banned user. You have to warn Google about your research, in which you may alter the results.

Weltevrede. I agree, I want to add that you can study the inner workings by looking at output, you can tell a lot about that

Feuz: it is an attempt to look at temporal relations: You have to try and fund ways to be able to ask these questions.

Society of the Query

Moderator; What I do not understand is the connection between the most opaque company ever which are still fetishing transparency.

Vaidhyanathan: it does not fetishize it; it leverages it. We do the work for Google, we provide the information and content Marx would scream at this notion. We are all very happy to do it (user-generated content). It is a better environment than we used to. However, we have to grasp the workings. Maybe we are very content with our relation to Google.

Weltevrede: it is also what building tools you can get out of Google. You can make use of the giant – building on Google; let Google work for us again.

Manovich (audience): I have difficulty to see your (Feuz’s and Weltevrede’s) results as research. What is the sample size? Where is the statistical data? You haven’t looked at the interdependencies of the variables? So what kind of science is this? If these things are not clear, these results are not meaningful.

Feuz: there is a difference between types of research. In the kind of research I did, I worked 4 month in a team gathering data. The amount of data we needed was already overwhelmingly large. You have to keep in mind that the thing is really temporal.

Vaidhyanathan (at Manovich): Is it not very expensive what you do? How can you do this?

Manovich: Most things are done in open source software and only takes five minutes.

Rogers (audience): Responds to the question by Manovich on what kind of science this is: it is diagnostics! Are local Googles furnishing local sources? It is a kind of critical diagnostics to see how Google works, and to see at the implications.

Manovich: Is it then issue exploration to be followed by hypothesis?

Moderator: I live in Silicon Valley, There is more skepticism there about Google. They cannot fight the real-time twitter economy. What is the relevancy of Google right now? What are your thoughts about this? Will it remains strong?

Vaidhyanathan: I am very bad at predicting. For the sake of my book, I hope they stay relevant? The rapid changes of Google have made me realize I must not write about current companies anymore. You have to keep in mind, though, that the real time web is not that connected (yet). So much of what Google tries to do is to satisfy the Cosmo-elite because this group makes the choices and the critics. What are the initiatives that Google has in India, China and Brazil? That is a more relevant development to look into.

Feuz; we researchers cannot cope with the patterns of change – they can adopt fast, so they will survive.

Society of the Query

Esther Weltevrede: “The Globalisation Machine. Reinterpreting engine results”

Posted: November 16, 2009 at 12:07 am  |  By: admin  |  Tags: , ,

Society of the Query

Lecture by Esther Weltevrede
As a follow up on Martin’s talk, I am going to present some empirical works. These project concern comparing engine results and customization of location. The aim of this study is:

1) Building on Googlization theory and search engine critique.
2) Empirical study. Reinterpreting Google for localization studies.

The key question is: What type of globalization machine is Google?
In this light, a couple of cases will be presented. Weltevrede starts by posing that PageRank is Googles way into the information environment. In an article published in 1998/1999 (?) PageRank is mentioned as the global ranking system for all pages, specifically designed for all the info of the world. Although Google states that they use a large number of variables, PageRank is primarily based on the link. The question of this research is: When Google moves to the local, what happens to the local results? What if we look at some services that are local:

A case:
Google “Amsterdam” and you get (respectively) red light, airport, coffee shops. This same query in Google.nl returns another set of results (arena, forest, tournament). Local domain Google is another method of localization (e.g. Google.de). There are 157 local Googles. The key variables are (as far as can be distilled: Google is not very transparent in providing this information):

  • IP address
  • top level domain
  • webmasters page

If you visit one of these Googles (say, Google.be), you can also select pages from that locale (only provide me with result from Belgium, for instance). If you select this option, you get local results according to Google. Also, we notice that this particular Google is recommended in three languages (French, Dutch and German, in this case). Another way that Google returns local results is via ‘region” and of course a yellow-page kind of search is offered via Google maps. In answering what we can say about the type of machine Google is, Weltevrede states that it thinks globally and acts locally.

The first case study:
Local and international information sources. Research question: to what extend can the local domain Google present local results? Method used: query all the national Googles in their official languages. Then, in Google Translate, the search term is translated. The second step is to geo- locate sources. Instead of choosing for host, we looked at registration of the website. This is a more precise indication of who owns the website. The top ten results for the query “human rights”.

A picture is shows about the results. The selected national Google is Canada:

map canada

This map indicates that Canada has relatively many local results for ‘human rights’. We can also look at what the top results are globally. The UN is by far the largest source in the list. When we looked at the results more closely, the declaration of human rights keeps popping up. Often websites have translated the declaration in all languages they all call upon this source (On e can interpret this as a way of SEO) .

Next, a ranked tag cloud is shown.
weltevrede_hr_tagCloud
We looked at these sources and blue- tagged sources contain the declaration of human rights. Next, a rank list of all countries queried is given. 40 % of all national Googles do not get any local results. If you look at the type list, you see that Europe leads the list, while at the lower end it is mostly African and Middle- Eastern countries. We can see that the local domain does not mean that you receive local information sources. How then are local results defined? Is it maybe language? A search is done on all Arabic countries. This shows a language web – a shared language space. Does that mean that there are no local sources? In Lebanon, the term “human rights” again is queried. While this does return results, these results do not make it to the top. Local sources are on the second page and beyond.

In order to test this claim (language) we looked at a geographical regions defined by languages: Europe is chosen due to its local and distinct languages. The visual below shows they have very local sources (Again indicated by black domain names). The EU- Googles hardly share sources – characterized by their local sources. This can be argued as a being a language web.

weltevrede_hr_tagcloud_eu

We now move to the last example: comparing two Portuguese speaking sources. Portugal compared to Brazil: Here we might conclude that the established local sources are privileged. Language webs prefer over local webs.

weltevrede_hr_Brazil_mappa weltevrede_hr_Portugal_mappa

Search engine literacy (2nd case study)
We can use Google to read society; we have a particular way of interpreting search engine results. One example method: reading domain names and their main issues. Again, the example of human rights is used here. If we query this, we see a very established result list, where sources are relatively fixed. What happens when we query for a more dynamic topic? In this case a query is done on RFID in 2004;  back then, this was a young space. We see sources competing for the top. Compared to the human rights space, it has rather young and technology-minded sites; the build- up of the list is really different. Another method for research is to look at issues returned:

weltevrede_rfid_sources

A third case study:
A query for “rights’ is performed. What is the top ten list of rights per country? The total list as shown. This research required reading and interpreting languages by the team members. The top ten of prominent rights in local domains were collected and visualized. The total image is shown.
weltevrede_rights_visual
The color code – blue rights are shared, while the black ones are culturally specific for domains.

If we zoom in, we see that in Italy, the unique rights are the right to forger and the right to nationality. In Japan, they have computer programming rights, for instance. In Australia, you have specifically man’s rights One favorite: Finland’s every-mans right to freedom to roam in nature. If we are to draw conclusion from this case study, they would be: the globalizing machine can show the shared as well as the locally specific. Google is localizing, regional, nation and local, showing shared and specific. Local results does not mean local sources. Also, different regions on the web are offered, mostly via language.

For more information, see Govcom.org and Digital Methods Initiative. DMI Project page on The Nationality of Issues. Repurposing Google for Internet Research.

Esther Weltevrede: "The Globalisation Machine. Reinterpreting engine results"

Posted: November 16, 2009 at 12:07 am  |  By: tjerk timan  |  Tags: , ,

Society of the Query

Lecture by Esther Weltevrede
As a follow up on Martin’s talk, I am going to present some empirical works. These project concern comparing engine results and customization of location. The aim of this study is:

1) Building on Googlization theory and search engine critique.
2) Empirical study. Reinterpreting Google for localization studies.

The key question is: What type of globalization machine is Google?
In this light, a couple of cases will be presented. Weltevrede starts by posing that PageRank is Googles way into the information environment. In an article published in 1998/1999 (?) PageRank is mentioned as the global ranking system for all pages, specifically designed for all the info of the world. Although Google states that they use a large number of variables, PageRank is primarily based on the link. The question of this research is: When Google moves to the local, what happens to the local results? What if we look at some services that are local:

A case:
Google “Amsterdam” and you get (respectively) red light, airport, coffee shops. This same query in Google.nl returns another set of results (arena, forest, tournament). Local domain Google is another method of localization (e.g. Google.de). There are 157 local Googles. The key variables are (as far as can be distilled: Google is not very transparent in providing this information):

  • IP address
  • top level domain
  • webmasters page

If you visit one of these Googles (say, Google.be), you can also select pages from that locale (only provide me with result from Belgium, for instance). If you select this option, you get local results according to Google. Also, we notice that this particular Google is recommended in three languages (French, Dutch and German, in this case). Another way that Google returns local results is via ‘region” and of course a yellow-page kind of search is offered via Google maps. In answering what we can say about the type of machine Google is, Weltevrede states that it thinks globally and acts locally.

The first case study:
Local and international information sources. Research question: to what extend can the local domain Google present local results? Method used: query all the national Googles in their official languages. Then, in Google Translate, the search term is translated. The second step is to geo- locate sources. Instead of choosing for host, we looked at registration of the website. This is a more precise indication of who owns the website. The top ten results for the query “human rights”.

A picture is shows about the results. The selected national Google is Canada:

map canada

This map indicates that Canada has relatively many local results for ‘human rights’. We can also look at what the top results are globally. The UN is by far the largest source in the list. When we looked at the results more closely, the declaration of human rights keeps popping up. Often websites have translated the declaration in all languages they all call upon this source (On e can interpret this as a way of SEO) .

Next, a ranked tag cloud is shown.
weltevrede_hr_tagCloud
We looked at these sources and blue- tagged sources contain the declaration of human rights. Next, a rank list of all countries queried is given. 40 % of all national Googles do not get any local results. If you look at the type list, you see that Europe leads the list, while at the lower end it is mostly African and Middle- Eastern countries. We can see that the local domain does not mean that you receive local information sources. How then are local results defined? Is it maybe language? A search is done on all Arabic countries. This shows a language web – a shared language space. Does that mean that there are no local sources? In Lebanon, the term “human rights” again is queried. While this does return results, these results do not make it to the top. Local sources are on the second page and beyond.

In order to test this claim (language) we looked at a geographical regions defined by languages: Europe is chosen due to its local and distinct languages. The visual below shows they have very local sources (Again indicated by black domain names). The EU- Googles hardly share sources – characterized by their local sources. This can be argued as a being a language web.

weltevrede_hr_tagcloud_eu

We now move to the last example: comparing two Portuguese speaking sources. Portugal compared to Brazil: Here we might conclude that the established local sources are privileged. Language webs prefer over local webs.

weltevrede_hr_Brazil_mappa weltevrede_hr_Portugal_mappa

Search engine literacy (2nd case study)
We can use Google to read society; we have a particular way of interpreting search engine results. One example method: reading domain names and their main issues. Again, the example of human rights is used here. If we query this, we see a very established result list, where sources are relatively fixed. What happens when we query for a more dynamic topic? In this case a query is done on RFID in 2004;  back then, this was a young space. We see sources competing for the top. Compared to the human rights space, it has rather young and technology-minded sites; the build- up of the list is really different. Another method for research is to look at issues returned:

weltevrede_rfid_sources

A third case study:
A query for “rights’ is performed. What is the top ten list of rights per country? The total list as shown. This research required reading and interpreting languages by the team members. The top ten of prominent rights in local domains were collected and visualized. The total image is shown.
weltevrede_rights_visual
The color code – blue rights are shared, while the black ones are culturally specific for domains.

If we zoom in, we see that in Italy, the unique rights are the right to forger and the right to nationality. In Japan, they have computer programming rights, for instance. In Australia, you have specifically man’s rights One favorite: Finland’s every-mans right to freedom to roam in nature. If we are to draw conclusion from this case study, they would be: the globalizing machine can show the shared as well as the locally specific. Google is localizing, regional, nation and local, showing shared and specific. Local results does not mean local sources. Also, different regions on the web are offered, mostly via language.

For more information, see Govcom.org and Digital Methods Initiative. DMI Project page on The Nationality of Issues. Repurposing Google for Internet Research.