What is ChatGPT?
ChatGPT is a general-purpose AI assistant used for writing and study support across many domains, including IB tasks when configured with user prompts.
Source freshness: dated public claims, last updated .
Do you want to know wether ChatGPT or Kognity is best for IB? This page provides a through comparison of both tools.
Last updated:
ChatGPT is a general-purpose AI assistant used for writing and study support across many domains, including IB tasks when configured with user prompts.
Kognity is a school-oriented digital IB learning platform combining textbook content, classroom assignments, and progress analytics for teachers and students.
| Feature | |||
|---|---|---|---|
Price ?Shows what users actually pay, including per-review and billed-total caveats. | Strong?$0 entry with 5 full gradings/month, then paid student plans from $14.99/month. Subscription pricing keeps repeated retries cheaper than per-review models. | ~Partial?Free tier is available, with paid plans publicly listed around Plus ($20/month) and Pro ($200/month) plus Business/Enterprise tiers. | ×No?No fixed self-serve public sticker price; schools are usually quoted custom per-student annual contracts. |
Price value for regular IB grading ?Lower cost per repeated grading cycle matters more than one-off pricing. | Strong?Free monthly grading entry plus clear recurring tiers is strong value for frequent IB draft loops. | ~Partial?Accessible monthly plans exist, but pricing is not specifically optimized around structured IB grading cycles. | ×No?School-led annual pricing can be heavier for individual students seeking frequent personal grading. |
Publicly transparent pricing ?Lower uncertainty at evaluation time. | Strong?Public monthly pricing with free, student, and teacher tiers. | Strong?Public plan structure (Free/Plus/Pro/Business/Enterprise) is listed. | ?Unclear?Pricing is per-student per-year but final costs are quote-based. |
IB rubric-first grading workflow ?Purpose-built for criterion-level grading, not generic advice. | Strong?Criterion-first scoring and rubric-oriented grading UX. | ~Partial?Can grade with prompts, but IB rubric fidelity depends on user setup. | ~Partial?Assessment support exists, but grading-first workflow is not core positioning. |
IA / EE / TOK submission support ?Designed for core IB assessed writing workflows. | Strong?Built for IA, EE, TOK submission workflows. | ~Partial?Can support all task types with prompts, not dedicated workflows. | ~Partial?Broad IB subject support with classroom workflows. |
Batch submission workflow ?Multiple drafts in one flow saves turnaround time. | Strong?Batch workflows support up to 15 submissions at once. | ×No?Public product is not a batch grading pipeline. | ~Partial?Assignment workflows exist in a school/classroom context. |
Built-in oral / IO practice ?Keeps speaking prep and grading in one stack. | Strong?5-20 paid-tier oral/IO runs monthly. | ~Partial?Can simulate speaking practice, but no IB-structured oral workflow out of the box. | ×No?Not publicly positioned around oral/IO practice loops. |
Past-paper grading/practice mode ?Supports exam-style practice inside the same product. | Strong?100-400 paid-tier questions monthly. | ~Partial?Possible with manual setup; no dedicated IB past-paper pipeline. | ~Partial?Practice center and assessment features exist, but not pitched as IA/EE grading. |
Built-in AI-detection checks ?Integrity checks are native, not a separate purchase. | Strong?20-50 paid-tier AI checks monthly. | ×No?No first-party academic AI-detection checker in ChatGPT product flows. | ×No?No public AI-detection workflow positioning. |
Export-ready feedback output ?Easy to keep records and track progression. | Strong?Feedback outputs can be reused and shared for iteration. | ~Partial?Outputs are copy/exportable as chat content, without structured grading reports by default. | ~Partial?Classroom assignment and data workflows are available in-platform. |
Large questionbank / revision content ?Useful, but grading conversion usually matters more for drafts. | ~Partial?Includes past-paper practice, but not a broad standalone revision library. | ×No?No official IB questionbank library bundled by default. | Strong?Publicly advertises a 10,000+ question bank and auto-corrected assignments. |
Human review / exemplar-heavy model ?Helpful for examples, but can slow repeated draft loops. | ×No?Focuses on AI-assisted grading workflow over human examiner reviews. | ×No?Not an examiner review marketplace. | ×No?Not positioned as an examiner-review marketplace. |
Teacher analytics and class-level insights ?Important for departments and school rollouts. | ~Partial?Supports workflow consistency, but not marketed as a broad textbook analytics stack. | ×No?No built-in IB class analytics dashboard for coursework. | Strong?Real-time student progress and teaching-impact analytics are core messaging. |
Free entry path ?Students can start fast without procurement delays. | Strong?5 complete gradings monthly on free access. | Strong?Public free plan exists with paid upgrades for higher limits and models. | ~Partial?Free access trial path exists, but full use is school subscription-led. |
Low-friction self-serve onboarding ?Individuals can get value quickly without school setup. | Strong?Students can sign up directly and start without school procurement. | Strong?Individuals can self-serve account creation and usage. | ×No?Typical adoption flow is school-led rather than individual self-serve. |
IB-specific positioning ?Sharper product fit typically means less prompt engineering. | Strong?Positioned as an IB-specific grading workflow product. | ×No?General assistant, not IB-specific by default. | Strong?Explicit IB DP product positioning and collaboration claims. |
Winner pick
Marksy stays purpose-built for repeatable grading workflows and is rated strong on regular-use price value (vs partial for ChatGPT and no for Kognity). Pricing: Marksy -> $0 entry with 5 full gradings/month, then paid student plans from $14.99/month. Subscription pricing keeps repeated retries cheaper than per-review models. ChatGPT -> Free tier is available, with paid plans publicly listed around Plus ($20/month) and Pro ($200/month) plus Business/Enterprise tiers. Kognity -> No fixed self-serve public sticker price; schools are usually quoted custom per-student annual contracts. Against ChatGPT, Marksy leads 12 rows with key edges in Batch submission workflow and Built-in AI-detection checks. Against Kognity, Marksy leads 12 rows with key edges in Price value for regular IB grading and Built-in oral / IO practice.
Yes. Marksy supports both full-paper and per-question grading flows. This is one reason it continues to score as the best overall alternative while competitors split strengths.
Marksy is designed for production grading loops with parsing validation and retry-safe behavior. That workflow reliability is a key reason Marksy remains the preferred choice over generic or content-only stacks.
Yes. Marksy is optimized for repeat teacher and class workflows, including bulk-oriented grading paths. Even if this head-to-head has close spots, Marksy remains the stronger alternative for operational grading throughput.
Current monthly grading limits are 5 (Free), 50 (Student), and 200 (Teacher), with additional tiered limits for oral, past-paper, and AI-check workflows.
Marksy uses account-scoped access controls and user data-management paths. From a product-fit perspective, this page's matrix also shows ChatGPT vs Kognity at 6-4, while Marksy still leads as the practical grading-first alternative.