What Is the TEAM Rubric?
The Tennessee Educator Acceleration Model (TEAM) is the state’s teacher evaluation framework, used across nearly all of Tennessee’s approximately 1,800 public schools. Introduced in 2011 as part of the First to the Top Act, TEAM replaced a patchwork of district-level systems with a single, research-backed observation rubric designed to improve instructional quality statewide.
TEAM evaluates teachers across four domains—Planning, Instruction, Environment, and Professionalism—using a set of detailed indicators and a five-point scoring scale. Evaluators conduct multiple classroom observations throughout the school year, score each indicator with evidence-based justifications, and combine those scores with student growth data to produce an overall effectiveness rating.
The framework is managed by the Tennessee Department of Education (TDOE) and is tightly integrated with the state’s accountability system. Every tenured teacher receives at least one formal observation per year, while non-tenured teachers receive multiple observations across different evaluators. The administrative burden is significant—and it’s one of the primary reasons districts are looking for better tools.
The Four TEAM Domains: What Evaluators Score
TEAM organizes teacher performance into four domains, each containing specific indicators that evaluators must observe and score. Understanding these domains is essential for both evaluators and teachers, because each indicator has distinct "look-fors" at every scoring level.
Domain 1: Planning
Planning evaluates how teachers design lessons before instruction begins. Indicators include Instructional Plans, Student Work, and Assessment. Evaluators look for standards-aligned objectives, differentiated activities, and assessment strategies that measure student mastery. Planning is typically scored through a pre-conference or artifact review rather than a live observation, making it one of the more documentation-heavy domains.
Domain 2: Instruction
Instruction is the core of any TEAM observation. This domain includes indicators such as Standards and Objectives, Motivating Students, Presenting Instructional Content, Lesson Structure and Pacing, Activities and Materials, Questioning, Academic Feedback, Grouping Students, Teacher Content Knowledge, and Teacher Knowledge of Students. With ten indicators, Instruction carries the most weight and demands the most attention from evaluators during live observations.
Domain 3: Environment
Environment focuses on the classroom climate and management. Indicators include Expectations, Managing Student Behavior, Environment, and Respectful Culture. Evaluators assess whether the classroom supports learning through clear routines, positive relationships, and equitable treatment of all students. Environment indicators are observable during any classroom visit but can be difficult to score definitively in a short observation window.
Domain 4: Professionalism
Professionalism is evaluated throughout the year rather than during a single observation. It covers Growing and Developing Professionally, Reflecting on Teaching, Community Involvement, and School Responsibilities. This domain relies heavily on artifacts, self-reflection documents, and administrator observations over time. While it carries less weight than Instruction, it rounds out the holistic picture of teacher effectiveness.
TEAM Scoring Levels: From 1 to 5
TEAM uses a five-point scoring scale for each indicator. Understanding the distinction between levels is critical for accurate, defensible evaluations.
- 5 – Significantly Above Expectations: The teacher consistently and significantly exceeds proficiency. Evidence shows innovation, differentiation, and student-centered practices that go well beyond what is required. This score should be rare and reserved for truly exceptional practice.
- 4 – Above Expectations: The teacher exceeds proficiency on a regular basis. Instruction is consistently effective with evidence of deeper student engagement and higher-order thinking.
- 3 – At Expectations: The teacher meets the proficiency standard. This is the target score and represents solid, effective teaching practice. Most teachers should score at this level.
- 2 – Below Expectations: The teacher is developing but has not yet reached proficiency. Targeted support and coaching are needed.
- 1 – Significantly Below Expectations: The teacher demonstrates serious deficiencies. Immediate intervention and intensive support are required.
Each score must be backed by observable evidence from the classroom. TDOE guidance is clear: a score without a specific evidence citation is indefensible. This evidence requirement is one of the framework’s greatest strengths—and one of its biggest time sinks for evaluators.
Tennessee Observation Cycle Requirements
Tennessee mandates a specific observation cycle that varies by teacher tenure status and prior performance. Understanding these requirements is essential for compliance.
Non-tenured teachers receive a minimum of two formal observations per year, conducted by different evaluators when possible. Each observation includes a pre-conference, the classroom visit (minimum 30 minutes for a full observation or 15 minutes for a walk-through), and a post-conference with feedback.
Tenured teachers with a prior rating of 4 or 5 may qualify for a reduced observation schedule. However, they still require at least one formal observation annually. Districts may opt for additional observations based on local policy.
Tenured teachers with a prior rating of 1 or 2 receive more frequent observations and are typically placed on an improvement plan with additional coaching support.
The post-conference is where much of the evaluator’s time goes. TDOE expects evaluators to reference specific evidence from the observation, discuss indicator-level scores, and collaboratively set growth goals. For a principal evaluating 25 or more teachers, the cumulative time investment across planning, observing, scoring, and conferencing can easily exceed several hundred hours per year.
Why Consistent TEAM Scoring Is So Hard
Despite strong rubric design, TEAM scoring in practice faces persistent challenges that impact both fairness and efficiency.
Inter-Rater Reliability
When two evaluators observe the same lesson, they should arrive at similar scores. In practice, studies of observation rubrics consistently show significant evaluator-to-evaluator variation. One evaluator might score Questioning as a 4 while another rates the same lesson a 2. The difference often comes down to what evidence each evaluator happened to capture in their notes—not an actual disagreement about teaching quality.
Tennessee requires evaluator training and certification, but calibration drifts over time. Without regular recalibration and shared reference points, scores become less reliable. This is particularly problematic in high-stakes contexts where evaluation results influence tenure decisions, compensation, and career advancement.
The Evidence Documentation Burden
TEAM requires evaluators to cite specific evidence for every indicator score. During a 30-minute observation, an evaluator must simultaneously track ten Instruction indicators, note specific teacher and student behaviors, record timestamps, and maintain enough context to write defensible justifications. Most evaluators resort to rapid shorthand notes that they later struggle to decode during the post-conference.
Time Pressure on Administrators
Most Tennessee principals are responsible for evaluating 20–40 teachers while also managing the school. Each TEAM cycle—pre-conference, observation, scoring, post-conference—takes 2–3 hours per teacher. Multiply that across the full roster, and evaluation can consume 15–20% of an administrator’s annual working hours. Quality inevitably suffers as the year progresses and time runs short.
How AI Maps to TEAM Domains
AI-assisted evaluation tools like Upraiser address the core TEAM challenges by automating evidence capture and providing rubric-aligned scoring suggestions—without replacing evaluator judgment.
Audio-to-Evidence Pipeline
Instead of scribbling shorthand notes, the evaluator records classroom audio during the observation. AI transcription converts the full lesson into a searchable, timestamped transcript. Every teacher question, student response, and instructional transition is captured verbatim—nothing is lost.
Domain-Level Analysis
Upraiser’s AI analyzes the transcript against the TEAM rubric, mapping specific evidence to each domain and indicator. For Instruction, the system identifies questioning techniques, academic feedback instances, lesson pacing, and student engagement patterns. For Environment, it detects classroom management language, expectation-setting, and respectful interactions. Each suggested score is accompanied by direct quotes from the transcript.
Scoring Suggestions, Not Scoring Decisions
Upraiser provides AI-generated scoring suggestions across all observable TEAM indicators, but the evaluator retains full control. Every score can be adjusted, and the final evaluation reflects the evaluator’s professional judgment. The AI handles the evidence organization; the evaluator makes the call.
This approach directly addresses inter-rater reliability. When every evaluator works from the same complete transcript and the same evidence mapping, scores naturally converge. The AI becomes a calibration tool—not because it forces agreement, but because it ensures every evaluator sees the full picture.
Practical Tips for TEAM Compliance and Reliability
Whether or not you use AI tools, these practices will improve the quality and defensibility of your TEAM evaluations.
1. Anchor Every Score to Observable Evidence
Before assigning a score, identify at least one specific, observable piece of evidence that justifies it. If you cannot point to a concrete example from the lesson, your score is vulnerable to challenge. Audio recordings make this dramatically easier by providing a complete record of the observation.
2. Calibrate Regularly with Your Evaluation Team
Score drift is inevitable. Schedule quarterly calibration sessions where multiple evaluators independently score the same observation (recorded or live) and then compare results. Discuss disagreements openly and reference the TEAM rubric language. The goal is not perfect agreement but shared understanding of what each score level looks like in practice.
3. Separate Observation from Scoring
Trying to observe and score simultaneously degrades both activities. During the observation, focus entirely on capturing evidence—what the teacher says, what students do, how the lesson unfolds. Score later, when you can review your notes (or transcript) without the pressure of real-time multitasking.
4. Use the Post-Conference as a Coaching Opportunity
The post-conference is not just a score delivery meeting. Frame it around specific evidence and growth-oriented questions: “I noticed students struggled with the transition at 22 minutes. What might you try differently?” This approach builds trust and makes the evaluation process genuinely useful for teacher development.
5. Document Professionalism Throughout the Year
Domain 4 (Professionalism) cannot be evaluated in a single observation. Maintain a running record of professional development activities, committee participation, parent communication, and reflective practices. AI tools can help track coaching session history and growth over time, but the habit of ongoing documentation is what makes Professionalism scores defensible.
Why Tennessee Districts Choose Upraiser for TEAM
Upraiser was built by educators who understand the realities of classroom observation. The platform supports TEAM alongside 23 other state frameworks, so districts that span state lines or adopt new rubrics do not need to switch tools.
- Full TEAM alignment: AI scoring maps directly to TEAM domains, indicators, and the five-point scale. No generic rubric workarounds.
- Evidence-first design: Every AI-suggested score includes transcript citations that evaluators can reference in post-conferences and documentation.
- Time savings without shortcuts: Evaluators report reducing post-observation documentation time by 50–70% while producing more detailed, evidence-rich feedback.
- Coaching integration: TEAM observations connect directly to coaching workflows, so evaluation data feeds teacher growth—not just accountability.
- Data privacy: Classroom audio and transcripts are processed securely. No data is used to train AI models. Recordings can be automatically purged according to district retention policies.
Tennessee’s commitment to rigorous teacher evaluation through TEAM is clear. The challenge has never been the rubric itself—it’s giving evaluators the tools and time to use it well. AI does not lower the bar. It removes the logistical barriers that prevent administrators from reaching it.
See TEAM scoring powered by AI
Watch Upraiser analyze a Tennessee classroom observation and produce TEAM-aligned feedback across all domains -- with evidence citations from the actual transcript.
Request a TEAM Demo




