[BACK]
Computers, Materials & Continua
DOI:10.32604/cmc.2022.031096
images
Article

Hyperparameter Tuned Deep Learning Enabled Cyberbullying Classification in Social Media

Mesfer Al Duhayyim1,*, Heba G. Mohamed2, Saud S. Alotaibi3, Hany Mahgoub4,5, Abdullah Mohamed6, Abdelwahed Motwakel7, Abu Sarwar Zamani7 and Mohamed I. Eldesouki8

1Department of Computer Science, College of Sciences and Humanities-Aflaj, Prince Sattam bin Abdulaziz University, Saudi Arabia
2Department of Electrical Engineering, College of Engineering, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh, 11671, Saudi Arabia
3Department of Information Systems, College of Computing and Information System, Umm Al-Qura University, Saudi Arabia
4Department of Computer Science, College of Science & Art at Mahayil, King Khalid University, Muhayel Aseer, Saudi Arabia
5Faculty of Computers and Information, Computer Science Department, Menoufia University, Egypt
6Research Centre, Future University in Egypt, New Cairo, 11845, Egypt
7Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam bin Abdulaziz University, Al-Kharj, Saudi Arabia
8Department of Information System, College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, Al Kharj, Saudi Arabia
*Corresponding Author: Mesfer Al Duhayyim. Email: m.alduhayyim@psau.edu.sa
Received: 10 April 2022; Accepted: 20 May 2022

Abstract: Cyberbullying (CB) is a challenging issue in social media and it becomes important to effectively identify the occurrence of CB. The recently developed deep learning (DL) models pave the way to design CB classifier models with maximum performance. At the same time, optimal hyperparameter tuning process plays a vital role to enhance overall results. This study introduces a Teacher Learning Genetic Optimization with Deep Learning Enabled Cyberbullying Classification (TLGODL-CBC) model in Social Media. The proposed TLGODL-CBC model intends to identify the existence and non-existence of CB in social media context. Initially, the input data is cleaned and pre-processed to make it compatible for further processing. Followed by, independent recurrent autoencoder (IRAE) model is utilized for the recognition and classification of CBs. Finally, the TLGO algorithm is used to optimally adjust the parameters related to the IRAE model and shows the novelty of the work. To assuring the improved outcomes of the TLGODL-CBC approach, a wide range of simulations are executed and the outcomes are investigated under several aspects. The simulation outcomes make sure the improvements of the TLGODL-CBC model over recent approaches.

Keywords: Social media; deep learning; cyberbullying; cybersecurity; hyperparameter optimization

1  Introduction

With the rising number of users on social media results in a new method of bullying. The latter term is termed as aggressive or intentional move that is performed by individuals or groups of persons by recurrently transmitting messages over time toward a sufferer who is not in a state to defend him or herself [1]. Bullying has become a portion of the community forever. With the emergence of the internet, it has been only a matter of time unless bullies discover their way on to this opportunistic and new medium. According to National Crime Prevention Council Cyberbullying (CB) is termed as the usage of mobile phones, Internet, or another device for sharing or posting messages or pictures that intentionally hurts or embarrasses any other individual. Several researchers state that amongst 10%–40% of internet consumers are considered the sufferer of CB [2,3]. Effects of CB have serious effects which may range from impermanent anxiety to suicide.

Identification of CB in mass communication is regarded as difficult job. Definition of what represents CB is completely subjective [4]. For instance, according to common people, recurrent usage of swear texts or messages may be regarded as bullying. But, teen-related mass media platforms namely Form spring are not considered bullying [5]. CBs attack victims on distinct subjects namely religion gender, and race. Based on the subjects of CB, glossary and perceived meaning of words change remarkably. CB awareness is raised in most countries because of the consequences described in this work. Correspondingly, most of the authors submitted their works by machine learning (ML) methods for identifying CB in an automatic way [6,7]. However, many studies in this domain have been carried out for the English language. Moreover, the study which is carried out usually utilizes text mining methods which are equivalent to the sentiment analysis study [8]. The reality is mass media post are context-dependent and interactive; therefore, it is not regarded as a standalone message. A lot of research work was conducted on explicit speech identification [9]. Still, more volume of studies becomes necessary for solving the implicit language which makes identifying cyber-bullying in mass communication a difficult task [10]. Nonetheless, advancement was made in the identification of cyber-bullying by methods of deep learning (DL) and ML. But many recent studies have to be advanced for offering an appropriate method which includes clear and indirect elements.

The authors in [11] presented XP-CB, a new cross-platform structure dependent upon Transformer and adversarial learning. The XP-CB is to improve a Transformer leveraging unlabelled data in the source and target platforms for appearing with a general representation but preventing platform-specific trained. The authors in [12] relate the performance of several words embedded approaches in fundamental word embedded approaches for present advanced language techniques for CB recognition. It can be utilized LightGBM and Logistic regression (LR) techniques for the classifier of bullying and non-bullying tweets. Murshed et al. [13] examine the hybrid DL technique named deep autoencoder with recurrent neural network (DEA-RNN), for detecting CB on Twitter social media networks. The presented DEA-RNN technique integrates Elman form RNN with an optimizing Dolphin Echolocation Algorithm (DEA) to fine-tune the Elman RNN parameter and decrease training time.

Dewani et al. [14] executed wide pre-processing on Roman Urdu micro-text. This classically includes development of Roman Urdu slang- phrase dictionary and mapping slang then tokenization. The unstructured data are more managed for handling encoder text format and metadata or non-linguistic feature. Also, it can be implemented wide experimental by applying convolution neural network (CNN), RNN-long short term memory (LSTM), and RNN-bidirectional LSTM (BiLSTM) techniques. Ahmed et al. [15] presented binary and multi-class classifier methods utilizing hybrid neural network (NN) for bully expression recognition in Bengali language. It can be utilized 44,001 users comment on popular public Facebook pages that fall as to 5 classes Religious, Sexual, Non-bully, Threat, and Troll.

This study introduces a Teacher Learning Genetic Optimization with Deep Learning Enabled Cyberbullying Classification (TLGODL-CBC) model in Social Media. The proposed TLGODL-CBC model intends to identify the existence and non-existence of CB in social media context. Initially, the input data is cleaned and pre-processed to make it compatible for further processing. Followed by, independent recurrent autoencoder (IRAE) model is utilized for the recognition and classification of CBs. Finally, the TLGO algorithm is used to optimally adjust the parameters related to the IRAE model. To assuring the improved outcomes of the TLGODL-CBC method, a wide range of simulations are executed and the outcomes are investigated in several aspects.

2  The Proposed Model

In this study, a new TLGODL-CBC model was established for identifying the existence and non-existence of CB in social media context. At the primary stage, the input data is cleaned and pre-processed to make it compatible for further processing. Then, the pre-processed data is fed into the IRAE method for the recognition and classification of CBs. Finally, the TLGO algorithm is used to optimally adjust the parameters related to the IRAE model and thus results in enhanced classification performance. Fig. 1 depicts the overall block diagram of TLGODL-CBC technique.

images

Figure 1: Overall block diagram of TLGODL-CBC technique

2.1 Data Pre-Processing

Primarily, the data was analyzed to a maximum degree of refinement by executing and subsequent the step stipulated under punctuation removal, Stop word remover, sentence segmentation, tokenization, and lower casing. These are the stages which are obtained to have data decreased to size, and so, it can be also eliminated unwanted data which is established from the data. During this method, it can be generated a generic pre-processed which resulted from the elimination of punctuation and is some non-letter character in every document. At last, the letter case of all the documents is lowered. The outcome of this method provided us with a sliced document text dependent upon n length with n-gram word-based tokenizer. The next stage then it had gone with the tokenization model is for transforming the token as to another normal format. Similar to stemming, the resolve of lemmatization is for minimizing inflectional procedures for a particular base procedure. Preposition and conjunction, Article, and some pronouns, i.e., are assumed to stop words.

2.2 IRAE Based Classification

At this stage, the pre-processed data is fed into the IRAE technique for the recognition and classification of CBs. The standard RNNs are inclined to gradient vanishing and explosion in the training because of the repeating multiplication of weighted matrix. IRNN is a novel recurrent network which efficiently resolves the gradient problem by altering the time-based gradient back propagation [1620]. Also, IRNN is employ unsaturated function (i.e., rectified linear unit (ReLU)) as an activation function and retains higher robustness and then training. The IRNN network is formulated as:

ht=σ(Wxt+uht1+b),(1)

In which the input weighted W is a matrix, the recurrent weighted u is a vector, and signifies the Hadamard products. All the neurons at time t are only linked to their individual neuron at time t1 as elementwise multiplication was implemented rather than vector multiplication. The single neuron is formulated as:

hn,t=σ(Wnxt+unhn,t1+bn),(2)

whereas Wn denotes the nth row of W and un implies the nth row of u matrix. The gradient of all the neurons is independently computed in back propagation as no communication occurs amongst the neuron from a similar layer. Considering the objective that minimizing at time step T is Jn, once the gradient was propagating back to step T, the computation outcomes are expressed as:

Jnhn,t=Jnhn,Thn,Tnhn,t=Jnhn,Tk=tT1hn,k+1hn,k=k=tT1σn.k+1 un=Jnδhn,TunTtk=tT1σn,k+1(3)

σntk+1 implies the derivation of activation function, and the derivation of ReLU activation function is both zero and one. If the gradient at time step T propagates to time step t, the minimal and maximal effectual gradient is considered that ϵ and γ correspondingly. If every neuron fulfills the criteria |μn|ϵ|[(Tt)ϵk=1T1σn,k+1,(Tt)γk=1T1σn,k+1] the gradient could not be vanishing or exploding at time t. In order to generally employed activation functions like rectified linear unit (ReLU) and tanh, its derivative is not superior to 1, for instance, |σntk+1|1|. Particularly to function ReLU utilized under this case, their derivative is either 0 or 1. IRAE integrates the features of IRNN and auto encoder (AE). Accordingly, IRAE is a recurrent connection infrastructure that allows it for training and transferring parameters besides the time direction. The single IRAE is also an unsupervised learning NN which efforts for reproducing the input by comparing the last output to input. The IRAE primary encoded the input xt based on Eq. (4) for obtaining the hidden expression ht next, it executes the decoded function on ht based on Eq. (5) for obtaining the reconstructed formulated yt.

ht=f(Wlxt+uht1+b),(4)

yt=g(W2ht+c),(5)

In which W1 refers the encoder weighted matrix, u signifies the recurrent weighted matrix, b denotes the encoded bias vector, W2 represents the decoded matrix, and c implies the decoder bias vector. For making the network sparse and enhance the trained speed, the ReLU function y=max\; {O,x} is chosen as the activation function during this case. Fig. 2 showcases the framework of RAE.

images

Figure 2: Structure of recurrent autoencoder

2.3 TLGO Based Parameter Optimization

Lastly, the TLGO algorithm is used to optimally adjust the parameters related to the IRAE model and thus results in enhanced classification performance. We have projected TLGO approach by integrating the feature of genetic algorithm (GA) with teaching learning based optimization (TLBO) [21]. Here, exploration and exploitation are 2 major features that should be considered. To accomplish superior outcome, there needs to be a balance among global and local searching agents. TLBO is well performed in exploitation phase, that is, finding the optimal solution in local searching space, and poorly performed in exploration phase. To resolve the imbalance between exploitation and exploration stage, we developed TLGO method by adding the mutation and crossover operators of GA into TLBO algorithm. GA performed well in exploration phase and has better convergence rate. At first, TLGO approach has similar step as TLBO step. The first two stages of TLBO approach that is., teacher and learner phases are added to TLGO method without making any changes. In learner stage, each learner interacts with others for improving their knowledge. For every learner X(i)g, X(r)g is selected arbitrarily (i r) and the population is updated using. The designated parent reproduced a novel offspring by carrying out the crossover operator. The mutation step can be used to arbitrarily invert a bit of chosen candidate and population can be upgraded when the mutation operation is implemented. Next, the fitness of population can be evaluated by using. The procedure repeats until end condition is satisfied.

The TLGO method uses exploration as well as exploitation phases and maintains a balance in local and global searching agents by using mutation and crossover operators of GA to TLBO technique. The presented method has good convergence rate with lesser computation effort, interms of finding an optimum solution. Algorithm 1 illustrates the working process of the presented TLGO approach.

images

3  Performance Validation

The experimental validation of the TLGODL-CBC model is tested using a benchmark dataset from Kaggle repository [22]. The dataset holds images under two class labels (Insult-1049/Normal-2898). Tab. 1 illustrates the dataset details.

images

The confusion matrices generated by the TLGODL-CBC model on distinct sizes of training (TR) and testing (TS) data are illustrated in Fig. 3. With 80% of TR data, the TLGODL-CBC model has recognized 743 samples into insult and 2212 samples into normal. Moreover, with 20% of TS data, the TLGODL-CBC approach has recognized 187 samples into insult and 548 samples into normal. Furthermore, with 70% of TR data, the TLGODL-CBC system has recognized 667 samples into insult and 1903 samples into normal. At last, with 30% of TS data, the TLGODL-CBC algorithm has recognized 282 samples into insult and 832 samples into normal.

imagesimages

Figure 3: Confusion matrix of TLGODL-CBC technique: a) 80% of TR, b) 80% of TS, c) 70% of TR, and d) 30% of TS

Tab. 2 reports a brief classifier result of the TLGODL-CBC model with 80% of TR data and 20% of TS data. The experimental results indicated that the TLGODL-CBC model has reached effectual outcomes in both cases. For instance, with 80% of TR data, the TLGODL-CBC model has offered average accuy, precn, recal, and Fmeasure of 93.60%, 91.59%, 92.09%, and 91.83% respectively. Also, with 20% of TS data, the TLGODL-CBC technique has obtainable average accuy, precn, recal, and Fmeasure of 93.04%, 91.02%, 91.38%, and 91.20% correspondingly.

images

Tab. 3 defines a detailed classifier result of the TLGODL-CBC model with 70% of TR data and 30% of TS data. The experimental outcomes outperformed that the TLGODL-CBC approach has reached effectual outcomes in both cases. For instance, with 70% of TR data, the TLGODL-CBC approach has accessible average accuy, precn, recal, and Fmeasure of 93.05%, 90.44%, 92.32%, and 91.31% correspondingly. Eventually, with 30% of TS data, the TLGODL-CBC model has offered average accuy, precn, recal, and Fmeasure of 94.01%, 92.07%, 91.67%, and 92.36% correspondingly.

images

Fig. 4 provides an average result analysis of the TLGODL-CBC model with varying TR/TS data. The figure reported that the TLGODL-CBC model has accomplished effectual outcomes under distinct sizes of TR/TS data.

imagesimages

Figure 4: Average analysis of TLGODL-CBC technique: a) 80% of TR, b) 80% of TS, c) 70% of TR, and d) 30% of TS

Fig. 5 offers the accuracy and loss graph analysis of the TLGODL-CBC method on distinct TR/TS datasets. The results exhibited that the accuracy value inclines to improve and loss value inclines to reduce with a higher epoch count. It can be also exposed that the training loss is lower and validation accuracy is higher on the test dataset.

images

Figure 5: Accuracy and loss analysis of TLGODL-CBC technique: a) 80:20 of TR/TS accuracy, b) 80:20 of TR/TS loss, c) 70:30 of TR/TS accuracy, and d) 70:30 of TR/TS loss

To assuring the enhanced performance of the TLGODL-CBC model, a comparison study is made with recent models in Tab. 4 [23]. Fig. 6 provides a comprehensive comparative precn examination of the TLGODL-CBC model with recent models. The figure implied that the RNN and support vector machine (SVM) models have resulted to lower precn values of 80.12% and 82.43% respectively. Along with that, the BLSTM and LSTM models have accomplished slightly enhanced precn values of 85.30% and 84.27% correspondingly. Along with that, the gated recurrent unit (GRU) and random forest (RF) models have accomplished reasonably precn values of 88.97% and 87.51% respectively. However, the TLGODL-CBC model has reached maximum precn of 92.07%.

images

images

Figure 6: Precn analysis of TLGODL-CBC technique with existing algorithms

Fig. 7 offers a detailed comparative recal analysis of the TLGODL-CBC approach with recent models. The figure implied that the RNN and SVM models have resulted to lower recal values of 83.69% and 82.38% correspondingly. Likewise, the BLSTM and LSTM models have accomplished slightly enhanced recal values of 89.93% and 83.69% respectively. Followed by, the GRU and RF techniques have accomplished reasonably recal values of 81.83% and 89.93% correspondingly. Finally, the TLGODL-CBC technique has reached maximum recal of 92.67%.

images

Figure 7: Recal analysis of TLGODL-CBC technique with existing algorithms

Fig. 8 illustrates a brief comparative accy investigation of the TLGODL-CBC system with recent models. The figure exposed that the RNN and SVM approaches have resulted in lesser accy values of 81.99% and 83.20% respectively. Likewise, the BLSTM and LSTM models have accomplished somewhat improved accy values of 89.50% and 88.51% correspondingly. In addition, the GRU and RF models have accomplished reasonably accy values of 82.77% and 83.84% correspondingly. Finally, the TLGODL-CBC model has reached maximum accy of 94.01%.

images

Figure 8: Accy analysis of TLGODL-CBC technique with existing algorithms

Fig. 9 demonstrates a comprehensive comparative F1measure analysis of the TLGODL-CBC technique with recent algorithms. The figure implied that the RNN and SVM models have resulted to lower F1measure values of 89.07% and 88.63% respectively. Also, the BLSTM and LSTM models have accomplished slightly enhanced F1measure values of 85.01% and 81.11% respectively. Next, the GRU and RF models have accomplished reasonably F1measure values of 81.73% and 82.01% correspondingly. At last, the TLGODL-CBC model has reached maximum F1measure of 92.36%. After observing the aforementioned tables and figures, it can be apparent that the TLGODL-CBC model has exhibited maximum performance over other methods.

images

Figure 9: F1measure analysis of TLGODL-CBC technique with existing algorithms

4  Conclusion

In this study, a novel TLGODL-CBC model was established for identifying the existence and non-existence of CB in social media context. At the primary stage, the input data is cleaned and pre-processed for making it compatible for further processing. Then, the pre-processed data is fed into the IRAE technique for the recognition and classification of CBs. Finally, the TLGO algorithm is used to optimally adjust the parameters related to the IRAE model and thus results in enhanced classification performance. To assuring the enhanced outcomes of the TLGODL-CBC technique, a wide range of simulations was carried out and the outcomes are investigated in several aspects. The simulation outcomes ensured the improvements of the TLGODL-CBC model over recent approaches. Thus, the TLGODL-CBC model has exhibited reasonable performance over other methods. In future, hybrid DL classification models can be included to boost the overall classification performance.

Funding Statement: The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work under Grant Number (RGP 2/46/43). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2022R140), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the Deanship of Scientific Research at Umm Al-Qura University for supporting this work by Grant Code: (22UQU4210118DSR12).

Conflicts of Interest: The authors declare that they have no conflicts of interest to report regarding the present study.

References

  1. V. Balakrishnan, S. Khan and H. R. Arabnia, “Improving cyberbullying detection using twitter users’ psychological features and machine learning,” Computers & Security, vol. 90, pp. 101710, 2020.
  2. A. Muneer and S. M. Fati, “A comparative analysis of machine learning techniques for cyberbullying detection on twitter,” Future Internet, vol. 12, no. 11, pp. 187, 2020.
  3. Z. L. Chia, M. Ptaszynski, F. Masui, G. Leliwa and M. Wroczynski, “Machine learning and feature engineering-based study into sarcasm and irony classification with application to cyberbullying detection,” Information Processing & Management, vol. 58, no. 4, pp. 102600, 2021.
  4. A. Ali and A. M. Syed, “Cyberbullying detection using machine learning,” Pakistan Journal of Engineering and Technology, vol. 3, no. 2, pp. 45–50, 2020.
  5. M. Arif, “A systematic review of machine learning algorithms in cyberbullying detection: Future directions and challenges,” Journal of Information Security and Cybercrimes Research, vol. 4, no. 1, pp. 1–26, 2021.
  6. A. Bozyiğit, S. Utku and E. Nasibov, “Cyberbullying detection: Utilizing social media features,” Expert Systems with Applications, vol. 179, pp. 115001, 2021.
  7. T. Alsubait and D. Alfageh, “Comparison of machine learning techniques for cyberbullying detection on youtube arabic comments,” International Journal of Computer Science & Network Security, vol. 21, no. 1, pp. 1–5, 2021.
  8. A. Kumar and N. Sachdeva, “Multimodal cyberbullying detection using capsule network with dynamic routing and deep convolutional neural network,” Multimedia Systems, 2021, https://doi.org/10.1007/s00530-020-00747-5.
  9. M. Alotaibi, B. Alotaibi and A. Razaque, “A multichannel deep learning framework for cyberbullying detection on social media,” Electronics, vol. 10, no. 21, pp. 2664, 2021.
  10. Y. Fang, S. Yang, B. Zhao and C. Huang, “Cyberbullying detection in social networks using bi-gru with self-attention mechanism,” Information, vol. 12, no. 4, pp. 171, 2021.
  11. P. Yi and A. Zubiaga, “Cyberbullying detection across social media platforms via platform-aware adversarial encoding,” arXiv:2204.00334 [cs], 2022, Accessed: Apr. 09, 2022. [Online]. Available: http://arxiv.org/abs/2204.00334.
  12. S. Pericherla and E. Ilavarasan, “Performance analysis of word embeddings for cyberbullying detection,” IOP Conference Series: Materials Science and Engineering, vol. 1085, no. 1, pp. 012008, 2021.
  13. B. A. H. Murshed, J. Abawajy, S. Mallappa, M. A. N. Saif and H. D. E. A. Ariki, “DEA-RNN: A hybrid deep learning approach for cyberbullying detection in twitter social media platform,” IEEE Access, vol. 10, pp. 25857–25871, 2022.
  14. A. Dewani, M. A. Memon and S. Bhatti, “Cyberbullying detection: Advanced preprocessing techniques & deep learning architecture for roman urdu data,” Journal of Big Data, vol. 8, no. 1, pp. 160, 2021.
  15. M. F. Ahmed, Z. Mahmud, Z. T. Biash, A. A. N. Ryen, A. Hossain et al., “Cyberbullying detection using deep neural network from social media comments in bangla language,” arXiv:2106.04506 [cs], 2021, Accessed: Apr. 09, 2022. [Online]. Available: http://arxiv.org/abs/2106.04506.
  16. L. Wang, R. Tao, H. Hu and Y. R. Zeng, “Effective wind power prediction using novel deep learning network: Stacked independently recurrent autoencoder,” Renewable Energy, vol. 164, pp. 642–655, 2021.
  17. A. Muthumari, J. Banumathi, S. Rajasekaran, P. Vijayakarthik, K. Shankar et al., “High security for de-duplicated big data using optimal simon cipher,” Computers, Materials & Continua, vol. 67, no. 2, pp. 1863–1879, 2021.
  18. R. Gopi, P. Muthusamy, P. Suresh, C. G. G. S. Kumar, I. V. Pustokhina et al., “Optimal confidential mechanisms in smart city healthcare,” Computers, Materials & Continua, vol. 70, no. 3, pp. 4883–4896, 2022.
  19. D. A. Pustokhin, I. V. Pustokhin, P. Rani, V. Kansal, M. Elhoseny et al., “Optimal deep learning approaches and healthcare big data analytics for mobile networks toward 5G,” Computers & Electrical Engineering, vol. 95, pp. 107376, 2021.
  20. J. A. Alzubi, O. A. Alzubi, M. Beseiso, A. K. Budati and K. Shankar, “Optimal multiple key-based homomorphic encryption with deep neural networks to secure medical data transmission and diagnosis,” Expert Systems, vol. 39, no. 4, pp. e12879, 2022.
  21. A. Manzoor, N. Javaid, I. Ullah, W. Abdul, A. Almogren et al., “An intelligent hybrid heuristic scheme for smart metering based demand side management in smart homes,” Energies, vol. 10, no. 9, pp. 1258, 2017.
  22. Dataset: https://www.kaggle.com/competitions/detecting-insults-in-socialcommentary/data?select=test.csv.
  23. C. Iwendi, G. Srivastava, S. Khan and P. K. R. Maddikunta, “Cyberbullying detection solutions based on deep learning architectures,” Multimedia Systems, 2020, https://doi.org/10.1007/s00530-020-00701-5.
images This work is licensed under a Creative Commons Attribution 4.0 International License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.