I have finally handed in my master thesis on Predicting Political Party Affiliation from Danish and Norwegian Parliamentary Speeches, and can now call myself MSc in Computer Science. By examining state-of-the-art NLP models against classical text classification methods, I show that party affiliation of a given speaker is indeed possible (best F1-scores=0.67). Predictions, however, depend on the diversity of the political ideologies of the parties, and on the methods used. State-of-the-art neural network methods (BERT) showed to perform on par or better than Support Vector Machines (trained on ‘contextual’ features), despite being trained on fewer words.
Thanks to Analyse & Tal for providing me with the Danish parliamentary dataset, as well as helping me run initial experiments. Check out their website kendditfolketing.dk, based on the same dataset, for a great overview of the intricacies and statistics on the Danish Parliament and politicians.
Please contact me if you are interested or if you want to read the thesis.
After the success of last years intership project booksby.ai in collaboration with ML artist Andreas Refsgaard, we have seen an increasing interest in our project. With features in several blogs and online media outlets, and shares across the planet, we thank everyone for the interest. Perhaps there is more to come in the future.
Through my internship last semester, me and Andreas Refsgaard created this fun project called Booksby.ai. Our initial thoughts was to make a website filled with AI-generated items to confuse the visitor and provoke status quo by showcasing the capabilities of state-of-the-art ML models.
After some brainstorming and playing around with different models and datasets, we decided to make an online bookshop, where ALL content would be generated by AI.
That is, author names, book titles, descriptions of the books, reviews, reviewers, book covers, the actual content and the prizes would be made by ML algorithms.
We ended up having a lot of fun and learning quite a lot about the different state-of-the-art text and image ML modelsout there.
Booksby.ai became an online bookstore which sells science fiction novels. As described on the website:
Through training, the artificial intelligence has been exposed to a large number of science fiction books and has learned to generate new ones that mimic the language, style and visual appearance of the books it has read.
None of the stories, titles, descriptions, book covers or reviews related to any of the books on Booksby.ai has been written or designed by humans.
So following my own successful installation on my Razer Blade 15 with NVIDIA GTX 1070 (Max Q), the word spread..
My friend invited me to help install Ubuntu, dual booted with Windows 10, on his MSI GT63 Titan 8RG with a sweet ass 1080 GTX GPU. So being confident as hell giving my own success, I thought it was easy peasy pie.. But no. Apparently for the 1080 it requires different drivers.
We had some issues trying to install tensorflow 1.8.0 GPU with CUDA 9.2 and cuDNN 7.1.4, which resulted in tensorflow gpu actually working fine, BUT the NVIDIA drivers installed along the way were not functioning on the GPU, which led to a 800×600 constant resolution. #lofi
So instead we had to go for CUDA 10.0, which seemed to work out fine for the NVIDIA 1080 GTX.
GANs need a whole lot of image data to output proper results. Therefore one should consider oneself lucky, if one comes upon a highly uniform and standardised image dataset for GAN training.
However if one is not that lucky, then a nice way to sort shitty image data from not-so-shitty image data, is to train a simple classifier on a subset of the images in the dataset.
So to sort good from bad, I trained a MobileNet model in Keras to distinguish between what I considered good images and bad images. After only 10 epochs, the model got 85 percent accuracy on the validation images. That could possibly get better, so I retrained a VGG16 to try.
I also increased the size of the dataset by flipping every single image vertically using PIL, so I now had a dataset double the original size (omg). Here is the script for that. After retraining the VGG16, I now got a 99 % accuracy on the validation set and I finally reached satisfaction.
And so I made a script that ran through the whole dataset, predicted the good images from the bad, and sorted the respective images in new folders. One with the good ones and one with the bad ones. From here, some rough mistakes of the model could quickly be moved into the right folder, and the GAN now have some pretty sweet uniform data to munch on.
So for the last couple of weeks I have been trying to retrain different models to try and improve my Keras know-how and also to be able to use my own models in other applications (such as object recognition live on the webcam), omg yes!
Of course along the way I had different struggles with overfitting and what not, but mostly when using larger models my session would crash. Over and over. I thought it was exhausting the CPU and not using the GPU, but after further investigation I came upon others on the world wide web, who faced a similar issue. My most pressing error would occur when loading in a model, using keras’ model.load() – and it seems it is due to a memory leak. So simply running keras.backend.clear_session() cleared up the memory and made the notebook run smoothly again. Check out the solution in this git thread.
A few months ago I first started experimenting with GANS. How amazing I thought. Andreas Refsgaard, who I am doing my internship in collaboration with, showed me this grand world of GANS. We wanted to start playing around with this crazy thing, so through Paperspace, I started running the tensorflow implementation of the DCGAN from this github repo.
This also was a fun exercise in finding and scraping the web for nice images to use as my training data. I found several ways of scraping images, but the straight away easiest one was Fatkun Batch Downloader, that works as a plugin in Chrome. From this I was able to download loads of images from Google and also pick and choose which was to drop or keep.
However this solution doesn’t work for training your GAN, since it needs standardised and indistinguishable data to actually output something that looks the same.
I ended up trying out with book covers, since they were easy to come by. I was using this site, which allowed me to download huge amounts of book covers in different sizes. One simple wget command through the terminal landed me 10000 images.
So after training a DCGAN on these images for 150 epochs I was left with the below output images, which I think looks pretty freaking kool.
Some overfitting is shown in the output, already after around 90 epochs this starts to happen. However… stay tuned, cuz next time we will see how Mikkel uses the Progressive Growing of GANS to train the book covers and achieve much better results.
So the proces of actually installing the NVIDIA driver that fit my GTX-1070 GPU was a waaay easier step than actually setting it up with tensorflow-gpu, for running machine learning models on the GPU instead of the CPU. After all this was the main point of getting the Razer Blade 15.
So after trying several online tutorials and getting error after freaking error, I uninstalled everything I tried to install already and started with this tutorial, which saved me.
Also this dude, copying the exact same tutorial, gave some pretty good hints how to avoid certain errors along the way.
I ran into more errors though. When trying to install tensorflow, at around step 13, a fun little error about some “NCCL-SLA.txt” introduced itself. This error occurred, apparently because that file didn’t exist. But fixed it by duplicating the “LICENCE.txt” file inside of /usr/local/cuda-9.2/targets/x86_64-linux and changing the duplicates name to “NCCL-SLA.txt”.
Afterwards when actually trying to run the pip install tensorflow*.whl, some “Permission Denied” error popped up. I later found out (after 1 hour on google) that I simply had to run the command with the --user flag like so: pip install tensorflow*.whl --user
So just got the new Razer Blade 15, and to set it up for both creative stuff (e.g. ableton) and machine learning I decided to make it dual boot between Ubuntu and Windows 10.
Googling around it looked like it could be trouble for the new Razer Blade to install Ubuntu, but following this tutorial on youtube it was actually pretty straight forward. A few glitches apeared, but managed to install it without errors.
After installing Ubuntu on 100 Gb partition on my C-drive I needed to set up the NVIDIA graphics drivers. One quick search gave me this tutorial, which easily helped me set up the system with the NVIDIA drivers. One thing to add though, is the fact that you need to specify “sudo apt install nvidia-drivers-390” instead of “nvidia-390”.
Now off to install CUDA and tensorflow-gpu to make sure future tensorflow models will run on my new kickass GPU.