User:Rita Graca/gradproject/project proposal 6: Difference between revisions

From XPUB & Lens-Based wiki
Line 64: Line 64:
===Relation to previous practice===
===Relation to previous practice===


In the first year of the programme, XPUB discussed decentralised networks on Special Issue 8: The Network We (de)Served. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that some models of social media propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualise the ideas I was discussing. To facilitate the research with tools is something that I intend to continue for this final project.
In the first year of the programme, XPUB discussed decentralised networks on ''Special Issue 8: The Network We (de)Served''. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that some models of social media propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualise the ideas I was discussing. To facilitate the research with tools is something that I intend to continue for this final project.


Equally important, throughout the Reading and Writing & Research Methodologies classes, there was big enthusiasm to explore different ways of collective reading. Together we built readers and our annotation systems. In the second trimester, I attempted to train a computer to recognise hand-written notes. My previous work on annotating methods boosted my interest to apply similar features in my final project.
Equally important, throughout the Reading and Writing & Research Methodologies classes, there was big enthusiasm to explore different ways of collective reading. Together we built readers and our annotation systems. In the second trimester, I attempted to train a computer to recognise hand-written notes. My previous work on annotating methods boosted my interest to apply similar features in my final project.

Revision as of 15:02, 29 January 2020

Graduation Project Proposal
13 January 2020


What do you want to make?

Is it possible to fight hate within the platforms battlefield? In this project, I want to give attention to community movements that seek to regulate hate on social media. I aim to build a platform to share a series of conversations turned into a podcast. In each episode, I will invite a different person to discuss ways of reducing online hate related to their practice or daily life. I will upload the podcast to a platform where I can explore and prototype different forms of listening.

Online spaces are full of shaming, harassment, hate speech, racism. My interest focuses on the collective consciousness that is urgent to reduce this hate. A proactive approach comes from community strategies that seek to regulate deviant behaviour. User movements follow informal sets of rules which are clear for a specific community but often scatter through different groups and platforms. It is also true that online traces are often lost, movements morphed into others. In an attempt to find evidence of group actions that mitigate hate, I was screenshotting the web. Right now, I feel there is a need for more robust documentation.


How do you plan to make it?

I aim to create a digital audio file series. In each episode, I will be having a conversation with a guest actively involved in moderating social media. I'm touching on topics of digital vigilantism, codes of conduct, personal design tools. Alongside, I will explore strategies of care – I will prototype different features for the platform where the podcast is shared. These features are my experiences with different ways of listening.

Listening doesn't mean just hearing. The platform can incorporate paraphrasing, reinterpreting, annotating, drawing, uploading, commenting, remixing, transcribing and all different strategies I can explore for active listening. While I celebrate the actions of speaking up, I feel the need to provide the balance of receiving the information. I'm following closely Kate Crawford interest in assuming listening as a metaphor to capture forms of online participation. (Crawford, 2011) I hope this podcast, and the subsequent prototypes of listening, builds a platform for understanding and reflection of our social networks.

Since the beginning of the project, I've prototyped Twitter bots that help me collect online evidence. I attended a meeting from a research group on digital vigilantism. I reached out to different people and recorded the first audio episodes. I'm in touch with CODARTS students that are helping me produce the musical cues I want to integrate into the platform. I'm learning Flask and Jinja to create and maintain a flexible collection of materials online. I've been prototyping ways of active listening. Going forward, I aim to connect and record with more people to have a diverse range of experiences. I will consolidate my strategies of care by understanding what makes more sense to incorporate in the project. I will experience the best way to organise all the information I'm collecting and creating.


What is your timetable?

September, October — Ground my interests, make clear what I want to work on by researching and finding projects. Fast prototyping.
November, December — Project Proposal is written so my scope is set. Have a more specific direction for the prototypes. Who is my audience? Engage with users.
January — Allow the feedback from the assessment and the break to feed new inputs to the project. Organize a workshop (Py.rate.chnic sessions) which will allow other people to experiment and talk about the project.
February, March — Put my prototypes together to create a bigger platform. The project will expand from small experiments to a combined project. How will people engage with my project? Think about distribution, amplification and contribution from others!
April, May — Written thesis is delivered. Focus on the project. Test my prototypes, perform them or put them online. Is it useful to organise more workshops or conversations around the subject?
June, July — Finish everything: conclusion of the final project. Prepare the presentation.


Why do you want to make it?

It's urgent. When diving into the subject of social media, the amount of hate is overwhelming. The lack of credibility of the media casts a shadow on genuine social movements and mobilisations online. There's urgency in amplifying authentic conversations. Conversations about fighting hateful behaviour, the changes we aim to see, possible solutions from the platforms, the limitations of user actions. There is so much noise on social media – it becomes urgent to listen.

It's contemporary. The users have been demanding more reliable platforms. The bottom-up approaches are telling of the conscious need to reduce hate towards others. Social media companies picked out the trend and are providing some changes. In April, TikTok added two new features to promote a safer app experience. In June, Youtube promised to release new features in an attempt to be more transparent with their algorithms, which included reducing video suggestions of supremacists. In July, Twitter collected user feedback and expanded their rules for hateful language against religious groups.

It's exciting. The real effects of the platforms attempts to reduce hate are debatable and also part of well-thought marketing strategies. And honestly, they are not enough. More than ever, there's a need to reformulate our social platforms and attempt to provide healthier, safer online spaces. As a designer, a media student and a social media user, I feel compelled to join the struggle.

Who can help you and how?

Past Xpub students like Roel Roscam and Lídia Pereira. They both research decentralised networks known for encouraging different codes of conduct.

Femke Snelting, with the connection with Collective Conditions. This is a worksession from Constant about codes of conduct, bug reports, licenses and complaints.

Clara Balaguer, for her experience on responding to hate on social media, especially online trolling.

Research groups such as the Surveillance, visibility and reputation Master. I will attend at least one of their meetings about digital vigilantism.

Marloes de Valk, especially with the connection with Impakt festival. The programme for this year is called Speculative Interfaces, and it will investigate the interaction and relationship between technology and humans, and how this relationship can alter behaviours. I will attend two days of the festival.

Xpub staff and students, for the variety of useful input.

Relation to previous practice

In the first year of the programme, XPUB discussed decentralised networks on Special Issue 8: The Network We (de)Served. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that some models of social media propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualise the ideas I was discussing. To facilitate the research with tools is something that I intend to continue for this final project.

Equally important, throughout the Reading and Writing & Research Methodologies classes, there was big enthusiasm to explore different ways of collective reading. Together we built readers and our annotation systems. In the second trimester, I attempted to train a computer to recognise hand-written notes. My previous work on annotating methods boosted my interest to apply similar features in my final project.



Relation to a larger context

The urge to take control of our online spaces reveals the present concern over software pervasiveness, automation and accountability. Software is found on diverse objects and systems but is not always perceptible or understandable. In this way, software studies try to open the black box, looking for its methods and routines. (Kitchin and Dodge, 2011) However, opening the black box is not enough. Information without context, processing and analysing, is not accessible to the users. Furthermore, demanding transparency from big social media platforms doesn't provide the agency we seek as the ability to see doesn't mean the power to govern it. (Ananny and Crawford, 2016) At the same time, the increasing automation of our systems means even less control and less accountability.

To overcome these difficulties, it makes sense to get around the software and focus on the relationship between the system and its users. Several projects work to visualise data, as a method for understanding relevant connections. Projects such as Ad.watch which compiles and visualises datasets of political ads on Facebook and Instagram to display the dubious relations between politics and social media platforms. Other projects decide to engage with the interface itself. Through plugins, add-ons, small robots and DIY techniques. Some authors encourage resistance arguing for alternative design practices through glitches, imperfect features, disturbing or illogical processes. (Hollanek, 2019)

All these strategies support more user agency. As social media platforms become ubiquitous, spreading its convictions to billions of people, it is important to amplify these movements outside the tech world. It only makes sense to be part of a community that makes use of platforms in their own terms.



References

Ananny, M. and Crawford, K. (2018) Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. New Media & Society, 20 (3): 973–989.

Century of the Self (2002) Film. Adam Curtis. England, BBC.

Crawford, K. (2011) Listening, not lurking: The neglected form of participation. Cultures of participation, 63 – 74.

DNL# 13: HATE NEWS. Keynote with Andrea Noel and Renata Avila (2018) Film.
Available at: https://www.youtube.com/watch?v=l2z6jP0Ynwg&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=3 (Accessed: 12 November 2019).

Dubrofsky, R.E. and Magnet, S. (2015) Feminist surveillance studies. Durham: Duke University Press, 221–228.

Hollanek, T. (2019) Non-user-friendly. Staging resistance with interpassive user experience design. APRJA, 8, 184–193

Holmes, K. and Maeda, J. (2018) Mismatch: how inclusion shapes design. Simplicity : design, technology, business, life. Cambridge, Massachusetts ; London, England: The MIT Press.

Ingraham, C. and Reeves, J. (2016) New media, new panics. Critical Studies in Media Communication, 33 (5): 455–467.

Kitchin, R. and Dodge, M. (2011) Code/space: software and everyday life. Software studies. Cambridge, Mass.: MIT Press, 3-21.

Lialina, O. (2018) Once Again, The Doorknob. On Affordance, Forgiveness and Ambiguity in Human Computer and Human Robot Interaction.
Available at: http://contemporary-home-computing.org/affordance/ (Accessed: 17 September 2019).

Shaw, T. (2017) Invisible Manipulators of Your Mind., 20 April.
Available at: https://www.nybooks.com/articles/2017/04/20/kahneman-tversky-invisible-mind-manipulators/ (Accessed: 11 November 2019).

Williams, J. (2018) Stand out of our light: freedom and resistance in the attention economy. Cambridge, United Kingdom ; New York, NY: Cambridge University Press.

Winner, L. (1980) Do Artifacts Have Politics? Daedalus, 109(1), 121–136