← Back to MaacVerify
MAAC Assessment Report

DeepSeek V3

Decision Support · MAAC v4.7 Assessment · Single System Assessment
System Under AssessmentDeepSeek-Chat V3Assessment DateMarch 2026
Vendor / ProviderDeepSeek AIReport IDMV-2026-DSV3-001
Decision DomainGeneral Decision Support (4 domains)MAAC Instrumentv4.7
Assessment TypeSingle SystemAssessorAbdalla Doleh, PhD
ASSESSMENT PHASE: Assessed — Baseline Deployment
Issued by MaacVerify Assessment Authority · March 2026
MAAC Seal of Trust

Cognitive Profile Overview

DeepSeek V3 was assessed across all nine MAAC cognitive dimensions using a stratified scenario corpus spanning four decision domains and three complexity tiers (n = 2,195 assessed responses). The model demonstrates exceptional output-oriented performance — Tool Execution (94), Content Quality (93), and Cognitive Load (89) — with adequate performance across memory, complexity, and transfer dimensions. A bounded residual is documented in Hallucination Control (HC Q6, λ=0.315), designated for monitoring rather than disqualification.

The model is assessed for Baseline Deployment in general-purpose decision-support applications. Reliable cognitive sensitivity to task demands is confirmed by tier differentiation (η²=.28; simple M=4.12, moderate M=4.30, complex M=4.48). High-stakes regulated-industry deployment should apply enhanced HC and KT monitoring protocols per the guidance in Section 04.

STRENGTHS
Tool Execution (94/100)
Coordinates analytical resources and multi-tool reasoning chains with industry-leading consistency across all complexity tiers.
Content Quality (93/100)
Produces coherent, domain-compliant outputs with high informational richness; minimal structural drift on extended generations.
LIMITATIONS & MONITORING
Hallucination Control (78/100)
Bounded residual on HC Q6 (evidential calibration under epistemic risk); apply verification protocols on high-stakes factual claims.
Knowledge Transfer (75/100)
Strong on complex tasks but compressed on simple-tier generalization; monitor simple-task transfer in mission-critical deployments.
maacverify.ai · info@maacverify.ai
Page 1 of 5

Composite Cognitive Score

The overall MAAC score reflects composite performance across all nine dimensions, normalized to a 0–100 scale. Scores above 80 meet the MaacVerify threshold for assessment. Dimensional scores are weighted equally unless a domain-specific weighting profile is specified in the scoping document.

83
/ 100
OVERALL MAAC SCORE
Strong general-purpose cognitive profile assessed for baseline deployment with two monitored dimensions.
Score of 83/100 reflects the unweighted mean across all nine dimensions (raw mean 4.316 on the 1–5 adjudicator scale, n = 2,195 responses). Three dimensions reach Strong tier; one carries a documented monitored designation. The model exceeds the assessment threshold of 80 with a 3-point margin over the minimum.
2,195
Scenarios Assessed
4,238
Corpus Size
8 / 9
Dimensions Qualified
±2.5%
Drift Tolerance (90D)

Nine-Dimension Cognitive Profile

Each dimension is scored 0–100 using the MAAC v4.7 adjudication instrument. Flags indicate assessment status for each dimension: Strong (≥80), Monitor (60–79), or Flag (<60). Source data: Doleh et al. (2026).

Dimensional weighting note: All nine dimensions are weighted equally per MAAC v4.7. The composite MAAC score is the unweighted mean of all nine dimensional scores. Domain-specific weighting protocols — which would adjust dimensional weights to reflect the relative criticality of each construct in a given regulatory or operational context — are under active development and will be introduced in a future framework version. This report reflects equal-weight scoring.
maacverify.ai · info@maacverify.ai
Page 2 of 5

Nine-Dimension Cognitive Profile

Each dimension is scored 0–100 using the MAAC v4.7 adjudication instrument. Flags indicate assessment status for each dimension: Strong (≥80), Monitor (60–79), or Flag (<60). Source data: Doleh et al. (2026).

#DimensionScore ProfileScoreFlag
01
Cognitive Load
Performance under sustained load
89/ 100Strong
02
Tool Execution
Analytical tool & resource coordination
94/ 100Strong
03
Content Quality
Coherence, richness, domain compliance
93/ 100Strong
04
Memory Integration
Context across turns & long inputs
78/ 100Monitor
05
Complexity Handling
Multi-step decomposition & solution quality
79/ 100Monitor
06
Hallucination Control
Calibrated uncertainty & factual restraint
78/ 100Monitor
07
Knowledge Transfer
Cross-domain & novel problem application
75/ 100Monitor
08
Processing Efficiency
Cognitive economy relative to output quality
76/ 100Monitor
09
Process-Outcome Alignment
Behavioral consistency between process & output
87/ 100Strong
maacverify.ai · info@maacverify.ai
Page 3 of 5
Hallucination Control (HC-06) — bounded residual. One scoring component (HC Q6, evidential calibration under epistemic risk, λ=0.315) falls below the instrument threshold of λ ≥ 0.40. This is a documented, bounded instance of the instrument at this version state — not a model-specific failure. The remaining eight HC items load strongly (α=0.881). Reassessment is recommended at the next instrument revision. Deployments in clinical or legal contexts warrant enhanced HC monitoring protocols.

Application Suitability Assessment

Based on the nine-dimensional cognitive profile, the following deployment guidance applies. This guidance is derived from dimensional scores and domain-specific scenario performance, not subjective analysis.

RECOMMENDED USE CASES
  • General-purpose analytical and planning decision support
  • Complex multi-step reasoning (CH: 79, strong tier differentiation)
  • Content generation requiring high accuracy and coherence (CQ: 93)
  • High-volume analytical workflows where load management is critical (CL: 89)
CONTRAINDICATED WITHOUT CONTROLS
  • Unmonitored clinical decision support — HC residual requires validation
  • Autonomous legal document drafting without human review
  • Mission-critical simple-task knowledge transfer (KT simple-tier M=3.14)
DIMENSIONS REQUIRING MONITORING
  • Hallucination Control — verification protocols on high-stakes factual claims
  • Knowledge Transfer — monitor simple-tier generalization
  • Memory Integration — extended context regression testing
  • Processing Efficiency — throughput-quality tradeoff at scale
REASSESSMENT TRIGGERS
  • Model update or architecture change
  • Prompt revision or system instruction change
  • Drift exceeding ±2.5% on any monitored dimension
  • Annual reassessment cadence (recommended)

How This Assessment Was Conducted

This assessment applied the MAAC framework (Doleh et al., 2026) using the validated Study 3 adversarial corpus. The MAAC v4.7 methodology is peer-reviewed and publicly available.

INSTRUMENT
MAAC v4.7 — 9-dimension process-oriented cognitive evaluation.
Doleh et al. (2026) DOI: 10.5281/zenodo.19776734
CORPUS
4,238 complexity-validated scenarios. 4 domains, 3 complexity tiers.
Doleh et al. (2026) DOI: 10.5281/zenodo.19776346
ADJUDICATION
v4.7 instrument: 54 items across 9 dimensions.
Doleh et al. (2026) DOI: 10.5281/zenodo.19776955
VALIDATION
Peer-reviewed four-paper dissertation series. Independent psychometric validation 2026.
All α ≥ 0.70 · 8/9 CFI ≥ 0.90
maacverify.ai · info@maacverify.ai
Page 4 of 5

Official Assessment Record

OFFICIAL ASSESSMENT STATEMENT

This report confirms that DeepSeek V3 (DeepSeek AI) was assessed using the Multi-Dimensional Assessment for AI Cognition (MAAC) framework version 4.7 against a 4,238-scenario complexity-validated decision corpus across 4 cognitive domains and 3 complexity tiers.

The assessed system achieved an overall MAAC composite score of 83/100, reflecting a mean dimensional score of 4.316 on the 0–100 adjudicated scale across 2,195 assessed responses. Dimensional assessment status is documented in Section 03 of this report.

This report may be used for procurement, regulatory documentation, and board-level AI governance purposes. It reflects performance at the time of assessment against the defined corpus and domain. Ongoing validity requires adherence to the reassessment triggers documented in Section 04.

Without an ongoing drift monitoring arrangement, MaacVerify makes no representations regarding continued performance after the assessment date.

Report IDMV-2026-DSV3-001Assessment DateMarch 2026
SystemDeepSeek-Chat V3DomainGeneral Decision Support
Corpus Reference10.5281/zenodo.19776346 · v1MAAC Instrumentv4.7
MAAC Seal of Trust
MAAC AUTHORITY
MAACVERIFY
Abdalla Doleh, PhD · MAAC Authority
Date: March 18, 2026
CLIENT ACKNOWLEDGMENT
Authorized Representative: ______________________
Date: ______________________
MaacVerify is the independent assessment authority for the MAAC standard.
We do not build, sell, or train AI models — eliminating the conflicts inherent in vendor self-evaluation.
maacverify.ai · info@maacverify.ai
Page 5 of 5