My Organised Chaos

The intention of this vlog is simply to walk you through my organized chaos when it comes to research and structuring my thoughts. While I have my moments where I heartily envy other peoples’ organisational process and structured thoughts, I think that, at the end of the day, we all have to work out a research strategy that works for ourselves – and ourselves specifically.

I have just started a postdoc at the Institute for Humanities in Africa at UCT, exploring the social impacts of artificial intelligence on health – hence the book you see in the video. For me, it is a slow process of decapsulating myself from my PhD work on dating apps, which had truly acquired the status of something I was living and evolving alongside with in my life. Weaning myself away from it towards a related-yet-different aspect of coded technologies is not something that happens overnight for me. Especially given that my thesis is based on ethnographic fieldwork and the stories it tells will never simply be filed away.

Similarly, wrapping my head around a new topic or vantage point takes time. I try to read all the things that tease my curiosity further, meaning I surround myself with loads of books, and virtual heaps of articles, blogs etc. Especially at the beginning stages, I struggle to explain my thoughts in a way that is coherent. There is simply too much I find interesting, and I need to take my time to simmer all the input down to a more tangible focus – the question that will help me navigate a complex social phenomenon. In the meantime, you will find that my notes are all over – in different word documents, in my various notebooks (which I’m very fond of; I need to be able to physically hold on to my thoughts at times!), sticky notes and so forth.

The compulsion to make triple sure everything has been noted down properly and securely probably also comes from a bad previous experience I had. Just before the submission deadline of my Bachelor of Social Science Honours thesis (many a moon ago), my laptop was stolen in a house break-in. This was before I knew about ‘clouds’. I didn’t have a proper internet connection at home either and only made occasional copies on my hard drive. Alas, never again will I rework an entire project based on some earlyish drafts last minute!

Back to the video. It only reveals the surface of what looks like pure madness. However, as time passes and I develop a clearer train of thought (for which, of course, I have a special notebook), things become much converged and my notes more cohesive. Perhaps not for everyone who might be able to decipher my handwriting, but for me. And I then start introducing discussions around the topics with people around me, which helps giving the developing story in my head a more distinct motif once again.

I do have moments when I’m fed up with all of this along the way. This is when I clean up my documents, make sure they are labelled and sorted in a way that still makes sense to me and that my latest notes are readily accessible to me. To-do lists always make me calmer, too. Every now and then, I’m trying out something new, like Microsoft OneNote, which I mentioned in the vlog. I do still need my handwritten notes in addition, though – they also have the advantage that one can cross off completed tasks off with real vigour.

Whatever the process that you have found to be working in your scholarly career thus far, I’d encourage you to focus on that and refine it in a way that makes sense for you. Value it, cultivate it, and – most importantly – acknowledge that there are times when nothing works and you just need to take a step back and do something fun!

The ones left behind

Last week, I cycled past a bus reading ‘5G – don’t get left behind’ on its back. This very bus drives through Cape Town’s city centre and its more affluent suburbs, but also transports many workers who come in from low-income areas. The message bothered me. It was there for to sell a product and thus not necessarily meant to convey a meaningful message. Still, it did echo assumptions that I find to be prominent in discussions on digital media and technological developments more broadly.

For one thing, there is the premise that there will be an improved humanity with an increasing access to information. Information flows tend to be almost religiously celebrated as having supreme value in and of themselves (also referred to as dataism), as being inherently progressive, and as levelling social playing fields.

Presenting technisation as a lofty ideal or a superior mode of being to achieve rather than something created from a particular vantage point effectively veils the authoritative regimes of the technological revolution we currently witness. This includes the cultures and values embedded in tech products. Very few women and people of colour are hired in tech industries, leading to the development of problematic algorithms.[1] Even more problematically, designs and codes are presented as neutral and gender- and colour-blind, much like the employment politics in bis tech.[2]

Adding to their opacity is the fact that tech products are often portrayed as independent actors. Power relations precipitating unequal access to resources that tie in with social, economic and educational developments are, consequently, neatly brushed under the discursive carpet. Framing access as a matter of capability and choice (reach it, grab it – or else get left behind) rather than something that forms part of a historical development supports the prioritisation of the needs of some while the experiences of others (those who cannot reach) are rendered even less visible and relevant for imagined futures.

In Cape Town, where the geographic, economic and social divisions of Apartheid are notoriously persistent, the ‘don’t get left behind’ paradigm seems particularly cynical. It foreshadows an even more unequal future and places the responsibility for ‘being left behind’ onto individuals unable, for example, to invest in 5G products. This form of exclusion severs itself from problematic histories of divisions and portrays the ones to come as both evitable (ones can make the “right” choices and catch up with tech) and as an inescapable future of insiders and outsiders – much like the narratives of numerous sci-fi plots.

Why sci-fi could be the secret weapon in China's soft-power arsenal |  Financial Times

It was throughout my studying Tinder that I grew increasingly intrigued by what lies behind the shiny, promising exteriors of technologies and artificial intelligence (AI). This is why I want to continue studying their impact on our well-being, social identities, politics, economies and demographic developments. Something I am very curious about is the role of algorithms in how we as their users come to understand ourselves, the world around us, and how we relate to others. I’m especially interested in the impacts of technologies on relationships of trust.

The more I read about AI more broadly, the more I find myself getting irritated with its overly positivistic representations. Especially when people like Amazon CO Jeff Bezos shamelessly flaunt their extraordinary wealth by taking a quick trip to space in a phallic-shaped rocket – and making some extra cash by selling spare seats to similarly wealthy people.

When products like the new Tesla humanoid robot named Optimus are developed and when Amazon’s AI assistant Alexa seems to have learned a little too much about your habits, it is useful to think back to Bezos’s phallus-shaped rocket – just as a memento of how the products we are sold as progressive are anything but neutral, nor are they necessarily designed for our needs. While there are well-intentioned inventions (especially in the medical field), AI and big tech should not be treated as inherently superior approaches to human sense-making but rather as complementing it if well-developed. This is because tech solutions are not “semi-sentient” as ultrarich AI-enthusiast Elon Musk promises his new human-replacement robot to be and they only have the “sense” of morality that has been encoded in them.

If left unchecked, the trajectory of dataism may very well be to the detriment of humanism. Thankfully, this is not a sci-fi movie or a zero-sum game. We are in a position in which we can still decide just how to handle these seemingly inevitable developments that are sprung on us from silicon-valley and co. We can contextualise and look at them as the political and socially momentous projects that they are. “Don’t get left behind” messages in this context should serve as a wake-up call. But instead of letting them induce panic and self-questioning as the advertisers appear to intend, we should treat it as a reminder to consider people at the margins and designing appropriate interventions instead of placing blame in the most inappropriate ways.

[1] For more on this, read D’ignazio, C. and Klein, L.F., 2020. Data feminism. MIT Press.

[2] See Noble, S.U., 2018. Algorithms of oppression. New York University Press.