Judging Criteria

Understand how your project will be evaluated by the HackArambh judging panel. Follow these criteria to maximize your chances of success.

Judging Criteria for HackArambh

Effective Date: October 25, 2025


HackArambh ("we," "us," or "our") operates the website https://hackarambh.com/ (the "Site"), a platform dedicated to fostering innovation through multi-competition events, including coding challenges, idea submissions, prototype developments, and industry-focused contests. These Judging Criteria ("Criteria") describe how submissions are evaluated in our competitions (each, a "Competition"). By participating, you agree to these Criteria, our Terms & Conditions, Submission Guidelines, and any Competition-specific rules.

Our judging process is fair, transparent, and designed to reward originality, technical excellence, and real-world impact. A panel of expert judges, mentors, and industry professionals reviews entries. Specific weights or emphases may vary by Competition—check the Competition page for details.

1. Overview of the Judging Process
Blind Review: Submissions are anonymized to ensure impartiality (e.g., no names or affiliations visible during initial scoring).

Stages:
- Initial Screening: Compliance check against Submission Guidelines.
- Scoring: Independent scoring by judges using the criteria below.
- Deliberation: Discussion for shortlisting and final winners.

Timeline: Scoring begins post-deadline; shortlist announced within 1-2 weeks; winners within 4 weeks.
Feedback: Shortlisted teams receive constructive feedback; all participants can request general insights.

Judges' decisions are final and not subject to appeal, though we reserve the right to adjust scores for fairness (e.g., due to technical issues).

2. Core Judging Criteria
Submissions are scored on a scale of 1-10 across five key areas, with total scores determining rankings. Weights are approximate and may be adjusted per Competition.



Criterion

Description

Weight

Key Evaluation Points

Innovation & Originality

How novel and creative is the idea or solution? Does it address a unique problem or approach?

30%

- Fresh perspectives on challenges.
- Avoidance of generic or copied concepts.
- Potential for disruption or new applications.

Technical Quality & Execution

Is the submission well-implemented, robust, and functional?

30%

- Code cleanliness, efficiency, and error-handling.
- Use of appropriate tools/technologies.
- Scalability and security considerations.

Feasibility & Impact

Can this be realistically developed and deployed? What is the potential societal, economic, or environmental benefit?

20%

- Practical roadmap and resource estimates.
- Measurable outcomes.
- Alignment with real-world needs.

Presentation & Clarity

How effectively is the submission communicated?

10%

- Clear documentation, visuals, and demos.
- Logical structure in pitches or videos.
- Accessibility for non-experts.

Adherence to Guidelines & Ethics

Does it follow rules, promote inclusivity, and demonstrate ethical practices?

10%

- Compliance with themes, formats, and deadlines.
- Responsible AI/use of data.
- Team collaboration and attribution.


Total Score Calculation: Weighted average of individual scores. Minimum passing score: 6/10 per criterion.

3. Criteria by Competition Type
While the core criteria apply universally, emphases differ:

3.1 Coding Challenges
Focus: Technical Quality (40% weight) and Innovation (25%).
Examples: Algorithm efficiency, edge-case handling, integration of APIs.

3.2 Idea Submissions
Focus: Innovation (35%) and Feasibility (25%).
Examples: Problem-solution fit, market validation, scalability potential.

3.3 Prototype Developments
Focus: Execution (35%) and Impact (25%).
Examples: User testing results, UI/UX design, prototype interactivity.

Hybrid Competitions blend these as specified.

4. Tie-Breakers and Additional Factors
Tie Resolution: Higher Innovation score breaks ties; if equal, judges' deliberation.
Diversity & Inclusion: Bonus consideration for underrepresented groups or inclusive designs (not scored, but noted).
Mentorship Engagement: Teams utilizing optional mentorship may receive holistic boosts.
Disqualifications: Automatic zero for plagiarism, incomplete entries, or ethical violations (detected via tools and manual review).

5. Judge Qualifications
Judges are selected for expertise (e.g., CTOs, professors, startup founders) and diversity. Conflicts of interest are disclosed and mitigated (e.g., recusal from related entries). Full panel bios are posted on the Competition page.

6. Post-Judging
Announcements: Winners publicized on the Site, social media, and via email. Prizes awarded per Terms & Conditions.
Usage Rights: Top submissions may be featured in showcases (with consent and credit).
Appeals: Limited to factual errors (e.g., scoring miscalculation); submit within 72 hours to judging@hackarambh.com.

7. Changes to These Criteria
We may update these Criteria to align with evolving standards or Competition needs. The latest version is posted here. Changes apply prospectively.

8. Contact Us
For questions about judging or feedback requests:
Email: hello@hackarambh.com
Support: hello@hackarambh.com (general queries)