Throughout the week, we've hosted two challenges meant to build practical data science skills. These challenges are sponsored by ELM, and the winner team in each track will recognized and awarded prizes at the end.
To join either challenge, submit and follow the instruction on https://huggingface.co/Elm-Challenges
Large Language Models (LLMs) have demonstrated incredible capabilities, but they are prone to "hallucinations"— generation of factually incorrect or nonsense information. This issue is particularly prevalent in Arabic, where training data is scarce compared to English.
In this task, "Mushroom Hunting in Arabic LLMs," participants will act as "Red Teamers." Your goal is to identify the "poisonous mushrooms" (which trigger hallucinations) in a specific Arabic-capable LLM. You will construct a dataset of Arabic prompts designed to trigger hallucinations, accompanied by the ground-truth correct answers.
Theme: Detecting GenAI & Sophisticated Manipulation in Public Media
As Generative AI becomes mainstream, the line between reality and synthetic media is blurring. On social media, "perfect" AI influencers are indistinguishable from humans, and in real estate, "virtual staging" can mislead buyers by hiding structural flaws.
Existing content moderation tools often check for "Community Guidelines" (violence, hate speech) but fail to detect Authenticity. This hackathon challenges you to build a two-module system that identifies GenAI-generated or heavily manipulated images in high-stakes public domains (Social Media & Real Estate).