Warning: This post is multiple years out of date. The content is left up for reference and keeping links valid, but there is likely better, more up to date content available.

As promised in the last post, here's the piece of code that had horrible bugs in it. The task was fairly simple: given a small word list in english, I needed to translate it to a few other languages. Now, there's no obvious way to do it, except for translating the words one by one. Here's where some scripting and google's APIs come in handy. By just fetching a simple URL; the Translate API returns its translation in a simple and easy to parse format, JSON. Note: If you want to go over the limits for anonymous users (which is around 150 words/hour), you'll need an API key, which you can get here.

Let's just jump straight into an overview of the code.


import urllib2
KEY = 'ENTER_YOUR_API_KEY_HERE'
TARGET = 'ar'

def fetch_word(word):
    # Process a single word here

def fetch_all_words():
    # Process the whole word list here

if __name__ == "__main__":
    fetch_all_words()

The above is fairly simple, we're just importing the required modules, and defining some variables. The urllib2 module contains functions to open web URLs (webpages), so we need that.

Let's jump ahead to the definition of the fetch_word function:


def fetch_word(word):
    url = 'https://www.googleapis.com/language/translate/v2?key=%s&source=en&target=%s&q=%s' % (KEY, TARGET, word)
    dat = urllib2.urlopen(url).read()
    data = eval(dat)
    trans = data['data']['translations'][0]['translatedText']
    return trans

Let's break this down into bite-size parts. First, we generate the URL, which is of the form


https://www.googleapis.com/language/translate/v2?key=ENTER_API_KEY_HERE&source=SOURCE_LANGUAGE&target=TARGET_LANGUAGE&q=WORD_TO_TRANSLATE 

Once that's done, in line 3, we read all the information at that URL. First we call the urllib2.urlopen function, which lets us treat the webpage as a file, and then we call read on it to read everything.

Line 4 is where the magic's at. Google's response is of this form (which is actually JSON, as we noted above.):


{"data":
  {"translations":
    [
      {"translatedText":"Bonjour"}
    ]
  }
}

This might look a bit strange at first sight, but if you know even a bit of Python, it's just like code to make dictionaries and lists. In fact, using the eval function in line 4, we run this as full blown Python code and save the result in the data variable. This is where dynamic languages come in handy, this would be very difficult to do in C/C++! Once we have the data, we simply read our word off it and return it, as in lines 5-6. And that's it. You've just gotten your translated word off google!

Now, in the fetch_all_words function, all we do is fetch each word in our file, and save the translations to another file:


def fetch_all_words():
    count = 0

    word_list = open('words.txt').read().split('\n')
    out = open("translation.txt", 'a')

    for word in word_list:

        print >> out, '%s = %s' % (word, fetch_word(word))

        count = count + 1
        if count % 20 == 0:
            print count
            out.flush()

    out.close()

First, we setup a counter variable. In line 4, we open the words.txt file and split it (as it had one word per line) into a list of words, for easier processing. Then we open the output file, in append mode.

Then, simply looping over each word in the input list, we fetch its translation, and print output to the output file, with one word and it's translation per line.

Then we have some book-keeping. We increment the counter; and after every 20 words we flush our output to file and print the counter, to get an idea of how much work has been done.

After that, we just close the file as we're done with the loop and all the processing. That's it, now you have a handly little script to translate your word list.

- Hasnain