🔤

Natural Language Processing

Stella Polaris Knowledge Center
Stella Polaris Knowledge Center

All reports

Name
Organization
Description
File
Tags
Year
Type
Link
Status
Self reference
Early Detection of Sexual Predators in Chats

Humboldt Universitat zu Berlin

A report on how deep learning transform models can classify grooming attempts. The authors of the report created a dataset that was then used by Viktor Bowallius and David Eklund in the report Grooming detection of chat segments using transformer models, where an f1 score of 0.98 was achieved.

Early_detection_of_sexual_predators_in_chat.pdf
Natural Language ProcessingClustering/Classification
2021
Research (peer reviewed)
aclanthology.org
Early Detection of Sexual Predators in Chats
Guarding the Guardians: Automated Analysis of Online Child Sexual Abuse

This paper presents an automated tool designed to analyze children's sexual abuse reports. By automating the analysis process of abuse complaints, the tool significantly reduces the risk of exposure to harmful content by categorizing the reports on three dimensions: Subject, Degree of Criminality, and Damage.

2308.03880v2.pdf
Criminal investigationNatural Language Processing
2023
Research (peer reviewed)
arxiv.org
Fine-Tuning Llama 2 Large Language Models for Detecting Online Sexual Predatory Chats and Abusive Texts

Cornell University

This paper proposes an approach to detection of online sexual predatory chats and abusive language using the open-source pretrained Llama 2 7B-parameter model, recently released by Meta GenAI. We fine-tune the LLM using datasets with different sizes, imbalance degrees, and languages (i.e., English, Roman Urdu and Urdu). Based on the power of LLMs, our approach is generic and automated without a manual search for a synergy between feature extraction and classifier design steps like conventional methods in this domain. Experimental results show a strong performance of the proposed approach, which performs proficiently and consistently across three distinct datasets with five sets of experiments. This study's outcomes indicate that the proposed method can be implemented in real-world applications (even with non-English languages) for flagging sexual predators, offensive or toxic content, hate speech, and discriminatory language in online discussions and comments to maintain respectful internet or digital communities.

2308.14683v1.pdf
Machine learningNatural Language Processing
2023
Research (peer reviewed)
arxiv.org