User:Rita Graca/gradproject/project proposal 6: Difference between revisions

From XPUB & Lens-Based wiki
No edit summary
 
(38 intermediate revisions by the same user not shown)
Line 1: Line 1:
'''<big>Graduation Project Proposal</big>''' <br>
'''<big><big>Project Proposal</big></big>''' <br>
13 January 2020
29 January 2020




Line 9: Line 9:


Online spaces are full of shaming, harassment, hate speech, racism. My interest focuses on the collective consciousness that is urgent to reduce this hate. A proactive approach comes from community strategies that seek to regulate deviant behaviour. User movements follow informal sets of rules which are clear for a specific community but often scatter through different groups and platforms. It is also true that online traces are often lost, movements morphed into others. In an attempt to find evidence of group actions that mitigate hate, I was screenshotting the web. Right now, I feel there is a need for more robust documentation.
Online spaces are full of shaming, harassment, hate speech, racism. My interest focuses on the collective consciousness that is urgent to reduce this hate. A proactive approach comes from community strategies that seek to regulate deviant behaviour. User movements follow informal sets of rules which are clear for a specific community but often scatter through different groups and platforms. It is also true that online traces are often lost, movements morphed into others. In an attempt to find evidence of group actions that mitigate hate, I was screenshotting the web. Right now, I feel there is a need for more robust documentation.




<gallery widths=350px heights=350px>
<gallery widths=350px heights=350px>
File:twittercancel.jpg | Example of ''cancel culture'' on Twitter. Important design features: hashtag, trends and retweet.
File:twittercancel.jpg | Users encouraging the boycott of the musician R. Kelly. Digital vigilantism allows users to denounce hateful content within the platform's structure.
File:witchesmastodon.png | Community rules for ''witches.live''. Another bottom-up strategy that has been receiving a lot of attention is the development of rigorous codes of conduct.  
</gallery>
</gallery>


===How do you plan to make it?===
===How do you plan to make it?===
Line 22: Line 23:
Listening doesn't mean just hearing. The platform can incorporate paraphrasing, reinterpreting, annotating, drawing, uploading, commenting, remixing, transcribing and all different strategies I can explore for active listening. While I celebrate the actions of speaking up, I feel the need to provide the balance of receiving the information. I'm following closely Kate Crawford interest in assuming listening as a metaphor to capture forms of online participation. (Crawford, 2011) I hope this podcast, and the subsequent prototypes of listening, builds a platform for understanding and reflection of our social networks.
Listening doesn't mean just hearing. The platform can incorporate paraphrasing, reinterpreting, annotating, drawing, uploading, commenting, remixing, transcribing and all different strategies I can explore for active listening. While I celebrate the actions of speaking up, I feel the need to provide the balance of receiving the information. I'm following closely Kate Crawford interest in assuming listening as a metaphor to capture forms of online participation. (Crawford, 2011) I hope this podcast, and the subsequent prototypes of listening, builds a platform for understanding and reflection of our social networks.


Since the beginning of the project, I've prototyped Twitter bots that help me collect online evidence. I attended a meeting from a research group on digital vigilantism. I reached out to different people and recorded the first audio episodes. I'm in touch with CODARTS students that are helping me produce the musical cues I want to integrate into the platform. I'm learning ''Flask'' and ''Jinja'' to create and maintain a flexible collection of materials online. I've been prototyping ways of active listening. Going forward, I aim to connect and record with more people to have a diverse range of experiences. I will consolidate my strategies of care by understanding what makes more sense to incorporate in the project. I will experience the best way to organise all the information I'm collecting and creating.
Since the beginning of the project, I've prototyped Twitter bots that help me collect online evidence. I attended a meeting from a research group on digital vigilantism. I reached out to different people and recorded the first audio episodes. I'm in touch with CODARTS students that are helping me produce the musical cues I want to integrate into the platform. I'm learning ''Flask'' and ''Jinja'' to create and maintain a flexible collection of materials online. I've been prototyping ways of active listening. I talked with moderators of different communities. Going forward, I aim to connect and record with more people to have a diverse range of experiences. I will consolidate my strategies of care by understanding what makes more sense to incorporate in the project. I will experience the best way to organise all the information I'm collecting and creating.




<gallery widths=350px heights=350px>
File:Bot timeline.png | Bot collecting traces of hashtag activism on Twitter.
</gallery>


<gallery widths=450px heights=350px>
File:annotated_eg.png | At the time of the Special Issue 9, a point of interest for everyone was annotations. I was curious if we could train a computer to see all of these traces, so I started prototyping some examples. Annotated example from data set.
</gallery>


===What is your timetable?===
===What is your timetable?===
Line 47: Line 48:


''It's exciting.'' The real effects of the platforms attempts to reduce hate are debatable and also part of well-thought marketing strategies. And honestly, they are not enough. More than ever, there's a need to reformulate our social platforms and attempt to provide healthier, safer online spaces. As a designer, a media student and a social media user, I feel compelled to join the struggle.
''It's exciting.'' The real effects of the platforms attempts to reduce hate are debatable and also part of well-thought marketing strategies. And honestly, they are not enough. More than ever, there's a need to reformulate our social platforms and attempt to provide healthier, safer online spaces. As a designer, a media student and a social media user, I feel compelled to join the struggle.


===Who can help you and how?===
===Who can help you and how?===


Marloes de Valk, because of her knowledge on persuasive design, especially with the connection with Impakt festival. The [https://impakt.nl/festival/2019-2/ Impakt festival] itself can stimulate my research. The programme for this year is called ''Speculative Interfaces'' and it will investigate the interaction and relationship between technology and humans, and how this relationship can alter behaviours. I will attend two days of the festival.
Past Xpub students like Roel Roscam and Lídia Pereira. They both research decentralised networks known for encouraging different codes of conduct.
Femke Snelting, with the connection with Collective Conditions. This is a worksession from Constant about codes of conduct, bug reports, licenses and complaints.


Past Xpub students like [https://luciadossin.net/ Lucia Dossin] and [http://ilu.servus.at/category/13-fed-up.html Lídia Pereira], which have work on similar topics.
Clara Balaguer, for her experience on responding to hate on social media, especially online trolling.


[http://www.janavirgin.com/ Joana Moll] and [http://snelting.domainepublic.net/ Femke Snelting], with the connection with ''Critical Interface Politics''. This year-long research is full of useful resources. Includes the ''Critical Interface Toolbox'' and also related to the ''[https://interfacemanifesto.hangar.org/index.php/Main_Page Interface Manifesto]''.
Research groups such as the Surveillance, visibility and reputation Master. I will attend at least one of their meetings about digital vigilantism.


Olia Lialina, an artist with a lot of work on web interfaces. I will be in her workshop in November, which explores early web pages through the interface of ''[https://oneterabyteofkilobyteage.tumblr.com/ The GeoCities archive]''. Approximately 382,000 home pages.
Marloes de Valk, especially with the connection with Impakt festival. The programme for this year is called Speculative Interfaces, and it will investigate the interaction and relationship between technology and humans, and how this relationship can alter behaviours. I will attend two days of the festival.


Xpub staff and students, for the variety of useful input.
Xpub staff and students, for the variety of useful input.
Line 63: Line 67:
===Relation to previous practice===
===Relation to previous practice===


I understand that design shapes our ability to access, participate in, and contribute to the world (Holmes, 2018). As a graphic designer myself, I always was interested in the biases I implement in the things I build. Especially when those determine who can engage and how they do it.
In the first year of the programme, XPUB discussed decentralised networks on ''Special Issue 8: The Network We (de)Served''. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that some models of social media propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualise the ideas I was discussing. To facilitate the research with tools is something that I intend to continue for this final project.


Last year we discussed decentralised networks on Special Issue 8: The Network We (de)Served, and this build up my interest in the subject. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that centralised models of technology propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualize the ideas I was discussing. To turn the research into tools is something that I intend to continue for this final project.  
Equally important, throughout the Reading and Writing & Research Methodologies classes, there was big enthusiasm to explore different ways of collective reading. Together we built readers and our annotation systems. In the second trimester, I attempted to train a computer to recognise hand-written notes. My previous work on annotating methods boosted my interest to apply similar features in my final project.




<gallery widths = 350px heights=350px>
<gallery widths=400px heights=300px>
File:Nothatbad1.png| Last year's work on how to publish networks
File:Nothatbad1.png| Last year's work on how to publish networks
File:Tool 1.png| Development of a tool to archive images in networks
File:Tool 1.png| Development of a tool to archive images in networks
File:annotated_eg.png | At the time of the Special Issue 9, a point of interest for everyone was annotations. I was curious if we could train a computer to see all of these traces, so I started prototyping some examples. Annotated example from data set.
</gallery>
</gallery>
<br clear=all>




===Relation to a larger context===
===Relation to a larger context===


The urge to take control of our online spaces reveals the present concern over software pervasiveness, automation and accountability. Software is found on diverse objects and systems but is not always perceptible or understandable. In this way, software studies try to open the black box, looking for its methods and routines. (Kitchin and Dodge, 2011) However, opening the black box is not enough. Information without context, processing and analysing, is not accessible to the users. Furthermore, demanding transparency from big social media platforms doesn't provide the agency we seek as the ability to see doesn't mean the power to govern it. (Ananny and Crawford, 2016) At the same time, the increasing automation of our systems means even less control and less accountability.
What can we do against violent social platforms filled with hate speech that harms users? A fair answer is to insist on responsibility either from the government, tech companies or international organisations. Laws, such as the NetzDG Law in Germany, are admirable initiatives. The NetzDG law (or the ''Netzwerkdurchsetzungsgesetz'') is controversial, but it aimed to give legal importance to flagging, complaining and reporting inside platforms. Not every country can rely on a democratic government. However, these laws can set an example for so many social media companies that are US based, as well as European data centres. These legal discussions deserve more encouragement.
 
To overcome these difficulties, it makes sense to get around the software and focus on the relationship between the system and its users. Several projects work to visualise data, as a method for understanding relevant connections. Projects such as Ad.watch which compiles and visualises datasets of political ads on Facebook and Instagram to display the dubious relations between politics and social media platforms. Other projects decide to engage with the interface itself. Through plugins, add-ons, small robots and DIY techniques. Some authors encourage resistance arguing for alternative design practices through glitches, imperfect features, disturbing or illogical processes. (Hollanek, 2019)


All these strategies support more user agency. As social media platforms become ubiquitous, spreading its convictions to billions of people, it is important to amplify these movements outside the tech world. It only makes sense to be part of a community that makes use of platforms in their own terms.  
Users have been putting more pressure on social platforms to act politically towards users, something that these businesses have been avoiding. In the US, publishers such as traditional newspapers curate content, so they have responsibility for what is published. US laws declare that an ''interactive computer service'' is not a publisher. (Communications Decency Act, 1996) This means computer services can't be held accountable for what their users publish. Facebook is a computer service. However, when it starts banning content and deciding what is appropriate content, it's making editorial decisions. There's still some confusion where lies the responsability to moderate hate content.


 
In 2016, four researchers analysed the policies of fifteen social media platforms. (Pater et al., 2016) At the time, all but Vine and VK, mention the term harassment in their policies. This means that it was already considered problematic behaviour. However, not one of the fifteen platforms was defining what constituted harassment. Is it publishing a personal address online, or is it a hurtful comment? Are there any repercussions for such behaviour? The urge to take control of our online spaces reveals the present concern over growing hateful platforms, the absence of safe places, and an overall lack of accountability. For these reasons, it's stimulating to understand how users are using their informal techniques to regulate hate on their social networks.
<gallery widths = 400px heights=400px>
File:adwatch.png| [https://ad.watch/ ad.watch] compiles and visualises datasets of political ads on Facebook and Instagram. This screenshot shows the activity of my city in Portugal. The political party that created controversial street billboards aimed at young people is also the biggest investor in social media advertising.
</gallery>
 
</div>




Line 95: Line 92:
Ananny, M. and Crawford, K. (2018) Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. ''New Media & Society'', 20 (3): 973–989.  
Ananny, M. and Crawford, K. (2018) Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. ''New Media & Society'', 20 (3): 973–989.  


''Century of the Self'' (2002) Film. Adam Curtis. England, BBC.
Constant (2019) ''Colective Conditions.''


Crawford, K. (2011) Listening, not lurking: The neglected form of participation. ''Cultures of participation'', 63 – 74.
Crawford, K. (2011) Listening, not lurking: The neglected form of participation. ''Cultures of participation'', 63 – 74.


''DNL# 13: HATE NEWS. Keynote with Andrea Noel and Renata Avila'' (2018) Film.<br>
''DNL# 13: HATE NEWS. Keynote with Andrea Noel and Renata Avila'' (2018) Film.<br>
Available at: https://www.youtube.com/watch?v=l2z6jP0Ynwg&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=3 (Accessed: 12 November 2019).
Available at: https://www.youtube.com/watch?v=l2z6jP0Ynwg&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=3 (Accessed: 12 November 2019).
''DNL# 13: HATE NEWS. Panel with Caroline Sinders, Øyvind Strømmen, Cathleen Berger, Margarita Tsomou'' (2018) Film. <br>
Available at: https://www.youtube.com/watch?v=9vTo_4kKqYM&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=6&t=0s (Accessed: 12 November 2019).


Dubrofsky, R.E. and Magnet, S. (2015) ''Feminist surveillance studies.'' Durham: Duke University Press, 221–228.
Dubrofsky, R.E. and Magnet, S. (2015) ''Feminist surveillance studies.'' Durham: Duke University Press, 221–228.


Hollanek, T. (2019) Non-user-friendly. Staging resistance with interpassive user experience design. ''APRJA'', 8, 184–193
Freeman, J. (2013) The Tyranny of Structurelessness. WSQ: Women’s Studies Quarterly, 41 (3–4): 231–246.


Holmes, K. and Maeda, J. (2018) ''Mismatch: how inclusion shapes design. Simplicity : design, technology, business, life.'' Cambridge, Massachusetts ; London, England: The MIT Press.
Holmes, K. and Maeda, J. (2018) ''Mismatch: how inclusion shapes design. Simplicity : design, technology, business, life.'' Cambridge, Massachusetts ; London, England: The MIT Press.
Line 110: Line 110:
Ingraham, C. and Reeves, J. (2016) New media, new panics. ''Critical Studies in Media Communication'', 33 (5): 455–467.
Ingraham, C. and Reeves, J. (2016) New media, new panics. ''Critical Studies in Media Communication'', 33 (5): 455–467.


Kitchin, R. and Dodge, M. (2011) ''Code/space: software and everyday life.'' Software studies. Cambridge, Mass.: MIT Press, 3-21.
Pater, J., Kim, M., Mynatt, E., et al. (2016) ''Characterizations of Online Harassment: Comparing Policies Across Social Media Platforms'', 369-374.


Lialina, O. (2018) ''Once Again, The Doorknob. On Affordance, Forgiveness and Ambiguity in Human Computer and Human Robot Interaction.'' <br>
Trottier, D. (2019) Denunciation and doxing: towards a conceptual model of digital vigilantism. ''Global Crime'', pp. 1–17.
Available at: http://contemporary-home-computing.org/affordance/ (Accessed: 17 September 2019).


Shaw, T. (2017) ''Invisible Manipulators of Your Mind.'', 20 April.<br>
Winner, L. (1980) Do Artifacts Have Politics? ''Daedalus'', 109(1), 121–136
Available at: https://www.nybooks.com/articles/2017/04/20/kahneman-tversky-invisible-mind-manipulators/ (Accessed: 11 November 2019).


Williams, J. (2018) ''Stand out of our light: freedom and resistance in the attention economy.'' Cambridge, United Kingdom ; New York, NY: Cambridge University Press.
</div>
 
Winner, L. (1980) Do Artifacts Have Politics? ''Daedalus'', 109(1), 121–136

Latest revision as of 17:51, 10 February 2020

Project Proposal
29 January 2020


What do you want to make?

Is it possible to fight hate within the platforms battlefield? In this project, I want to give attention to community movements that seek to regulate hate on social media. I aim to build a platform to share a series of conversations turned into a podcast. In each episode, I will invite a different person to discuss ways of reducing online hate related to their practice or daily life. I will upload the podcast to a platform where I can explore and prototype different forms of listening.

Online spaces are full of shaming, harassment, hate speech, racism. My interest focuses on the collective consciousness that is urgent to reduce this hate. A proactive approach comes from community strategies that seek to regulate deviant behaviour. User movements follow informal sets of rules which are clear for a specific community but often scatter through different groups and platforms. It is also true that online traces are often lost, movements morphed into others. In an attempt to find evidence of group actions that mitigate hate, I was screenshotting the web. Right now, I feel there is a need for more robust documentation.



How do you plan to make it?

I aim to create a digital audio file series. In each episode, I will be having a conversation with a guest actively involved in moderating social media. I'm touching on topics of digital vigilantism, codes of conduct, personal design tools. Alongside, I will explore strategies of care – I will prototype different features for the platform where the podcast is shared. These features are my experiences with different ways of listening.

Listening doesn't mean just hearing. The platform can incorporate paraphrasing, reinterpreting, annotating, drawing, uploading, commenting, remixing, transcribing and all different strategies I can explore for active listening. While I celebrate the actions of speaking up, I feel the need to provide the balance of receiving the information. I'm following closely Kate Crawford interest in assuming listening as a metaphor to capture forms of online participation. (Crawford, 2011) I hope this podcast, and the subsequent prototypes of listening, builds a platform for understanding and reflection of our social networks.

Since the beginning of the project, I've prototyped Twitter bots that help me collect online evidence. I attended a meeting from a research group on digital vigilantism. I reached out to different people and recorded the first audio episodes. I'm in touch with CODARTS students that are helping me produce the musical cues I want to integrate into the platform. I'm learning Flask and Jinja to create and maintain a flexible collection of materials online. I've been prototyping ways of active listening. I talked with moderators of different communities. Going forward, I aim to connect and record with more people to have a diverse range of experiences. I will consolidate my strategies of care by understanding what makes more sense to incorporate in the project. I will experience the best way to organise all the information I'm collecting and creating.



What is your timetable?

September, October — Ground my interests, make clear what I want to work on by researching and finding projects. Fast prototyping.
November, December — Project Proposal is written so my scope is set. Have a more specific direction for the prototypes. Who is my audience? Engage with users.
January — Allow the feedback from the assessment and the break to feed new inputs to the project. Organize a workshop (Py.rate.chnic sessions) which will allow other people to experiment and talk about the project.
February, March — Put my prototypes together to create a bigger platform. The project will expand from small experiments to a combined project. How will people engage with my project? Think about distribution, amplification and contribution from others!
April, May — Written thesis is delivered. Focus on the project. Test my prototypes, perform them or put them online. Is it useful to organise more workshops or conversations around the subject?
June, July — Finish everything: conclusion of the final project. Prepare the presentation.


Why do you want to make it?

It's urgent. When diving into the subject of social media, the amount of hate is overwhelming. The lack of credibility of the media casts a shadow on genuine social movements and mobilisations online. There's urgency in amplifying authentic conversations. Conversations about fighting hateful behaviour, the changes we aim to see, possible solutions from the platforms, the limitations of user actions. There is so much noise on social media – it becomes urgent to listen.

It's contemporary. The users have been demanding more reliable platforms. The bottom-up approaches are telling of the conscious need to reduce hate towards others. Social media companies picked out the trend and are providing some changes. In April, TikTok added two new features to promote a safer app experience. In June, Youtube promised to release new features in an attempt to be more transparent with their algorithms, which included reducing video suggestions of supremacists. In July, Twitter collected user feedback and expanded their rules for hateful language against religious groups.

It's exciting. The real effects of the platforms attempts to reduce hate are debatable and also part of well-thought marketing strategies. And honestly, they are not enough. More than ever, there's a need to reformulate our social platforms and attempt to provide healthier, safer online spaces. As a designer, a media student and a social media user, I feel compelled to join the struggle.


Who can help you and how?

Past Xpub students like Roel Roscam and Lídia Pereira. They both research decentralised networks known for encouraging different codes of conduct.

Femke Snelting, with the connection with Collective Conditions. This is a worksession from Constant about codes of conduct, bug reports, licenses and complaints.

Clara Balaguer, for her experience on responding to hate on social media, especially online trolling.

Research groups such as the Surveillance, visibility and reputation Master. I will attend at least one of their meetings about digital vigilantism.

Marloes de Valk, especially with the connection with Impakt festival. The programme for this year is called Speculative Interfaces, and it will investigate the interaction and relationship between technology and humans, and how this relationship can alter behaviours. I will attend two days of the festival.

Xpub staff and students, for the variety of useful input.


Relation to previous practice

In the first year of the programme, XPUB discussed decentralised networks on Special Issue 8: The Network We (de)Served. I understood how creating new platforms and looking for alternatives reveals the desire for bottom-up changes and more active end-users. It became clear that some models of social media propagate limited ideas, and those ideas shape our society. As a project for that same Special Issue, I prototyped some tools that helped me visualise the ideas I was discussing. To facilitate the research with tools is something that I intend to continue for this final project.

Equally important, throughout the Reading and Writing & Research Methodologies classes, there was big enthusiasm to explore different ways of collective reading. Together we built readers and our annotation systems. In the second trimester, I attempted to train a computer to recognise hand-written notes. My previous work on annotating methods boosted my interest to apply similar features in my final project.



Relation to a larger context

What can we do against violent social platforms filled with hate speech that harms users? A fair answer is to insist on responsibility either from the government, tech companies or international organisations. Laws, such as the NetzDG Law in Germany, are admirable initiatives. The NetzDG law (or the Netzwerkdurchsetzungsgesetz) is controversial, but it aimed to give legal importance to flagging, complaining and reporting inside platforms. Not every country can rely on a democratic government. However, these laws can set an example for so many social media companies that are US based, as well as European data centres. These legal discussions deserve more encouragement.

Users have been putting more pressure on social platforms to act politically towards users, something that these businesses have been avoiding. In the US, publishers such as traditional newspapers curate content, so they have responsibility for what is published. US laws declare that an interactive computer service is not a publisher. (Communications Decency Act, 1996) This means computer services can't be held accountable for what their users publish. Facebook is a computer service. However, when it starts banning content and deciding what is appropriate content, it's making editorial decisions. There's still some confusion where lies the responsability to moderate hate content.

In 2016, four researchers analysed the policies of fifteen social media platforms. (Pater et al., 2016) At the time, all but Vine and VK, mention the term harassment in their policies. This means that it was already considered problematic behaviour. However, not one of the fifteen platforms was defining what constituted harassment. Is it publishing a personal address online, or is it a hurtful comment? Are there any repercussions for such behaviour? The urge to take control of our online spaces reveals the present concern over growing hateful platforms, the absence of safe places, and an overall lack of accountability. For these reasons, it's stimulating to understand how users are using their informal techniques to regulate hate on their social networks.


References

Ananny, M. and Crawford, K. (2018) Seeing without knowing: Limitations of the transparency ideal and its application to algorithmic accountability. New Media & Society, 20 (3): 973–989.

Constant (2019) Colective Conditions.

Crawford, K. (2011) Listening, not lurking: The neglected form of participation. Cultures of participation, 63 – 74.

DNL# 13: HATE NEWS. Keynote with Andrea Noel and Renata Avila (2018) Film.
Available at: https://www.youtube.com/watch?v=l2z6jP0Ynwg&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=3 (Accessed: 12 November 2019).

DNL# 13: HATE NEWS. Panel with Caroline Sinders, Øyvind Strømmen, Cathleen Berger, Margarita Tsomou (2018) Film.
Available at: https://www.youtube.com/watch?v=9vTo_4kKqYM&list=PLmm_HP_Sb_cTFwQrgkRvP8yqJqerkttpm&index=6&t=0s (Accessed: 12 November 2019).

Dubrofsky, R.E. and Magnet, S. (2015) Feminist surveillance studies. Durham: Duke University Press, 221–228.

Freeman, J. (2013) The Tyranny of Structurelessness. WSQ: Women’s Studies Quarterly, 41 (3–4): 231–246.

Holmes, K. and Maeda, J. (2018) Mismatch: how inclusion shapes design. Simplicity : design, technology, business, life. Cambridge, Massachusetts ; London, England: The MIT Press.

Ingraham, C. and Reeves, J. (2016) New media, new panics. Critical Studies in Media Communication, 33 (5): 455–467.

Pater, J., Kim, M., Mynatt, E., et al. (2016) Characterizations of Online Harassment: Comparing Policies Across Social Media Platforms, 369-374.

Trottier, D. (2019) Denunciation and doxing: towards a conceptual model of digital vigilantism. Global Crime, pp. 1–17.

Winner, L. (1980) Do Artifacts Have Politics? Daedalus, 109(1), 121–136