Overview
HR Technology Conference 2025 returns to Las Vegas at Mandalay Bay in September with AI—especially agentic, workflow-embedded assistants—dominating the agenda and expo.
Buyers can expect SHRM and HRCI credits across selected sessions, a crowded Startup Pavilion, and a high-stakes Pitchfest finale designed to spotlight measurable ROI and enterprise readiness.
- Where and when: Mandalay Bay, Las Vegas, September 2025 (four days including pre-conference programs)
- Headlines: Agentic AI moves from pilots to production workflows; risk, governance, and skills intelligence mature
- Accreditation: SHRM and HRCI recertification credits are available on eligible sessions; claiming guidance below
- Pitchfest: Finalists span TA automation, skills intelligence, and EX co-pilots; winners recognized on the final day
- Buying signals: Expect usage-based AI pricing add-ons, deeper HRIS/ATS integrations, and more skills graph connectors
- What to do now: Use the 30-60-90 buyer playbooks, procurement guardrails, and integration checklists to shortlist and budget
Implication for buyers: Arrive with a focused use-case list and a governance-first stance—agentic AI and integrations will define 2025 value creation and risk posture.
What, where, who: dates, venue, and verified attendance and exhibitor stats
HR Technology Conference 2025 is slated for September at Mandalay Bay in Las Vegas. The show maintains its format of keynotes, an expansive expo, and startup programming.
Historically, official attendance and exhibitor totals are published by show organizers as the program concludes. Consult the HR Technology Conference site and news updates for final counts and category breakdowns.
The venue’s scale supports hundreds of exhibitors across TA, HRIS suites, payroll, analytics, and EX. Expect a dedicated Startup Pavilion and a Pitchfest stage.
Capacity often varies by program type (keynotes, breakout rooms, pre-conference workshops). “Sold out” typically applies to specific ticket types or workshops rather than the entire expo.
Implication: Secure hotel blocks and workshop seats early. Use verified post-show stats to benchmark vendor category density and plan 1:1 meetings.
SHRM/HRCI credits and eligible sessions at HR Tech 2025
You can earn SHRM and HRCI recertification credits from eligible HR Tech 2025 sessions. Credit types and quantities are labeled in session descriptions.
SHRM-CP/SHRM-SCP credits are recorded as Professional Development Credits (PDCs). HRCI credits may be General, Business (SPHR), or Ethics depending on the content.
To claim credits, confirm eligibility in the session description and capture session codes or confirmation scans. Submit via your certification portal. See SHRM recertification and HRCI recertification for requirements and documentation standards.
Keynotes are not always credit-bearing. Workshops frequently carry higher credit values.
Implication: Map your learning plan to your renewal window. Prioritize sessions that support your competency gaps (e.g., analytics, risk, or AI governance).
Registration and pricing intelligence: ticket tiers, discount windows, and sell-out patterns
Expect multi-tier ticketing—commonly Early Bird, Advance, and Standard/Onsite. Workshops, Investor Experience, and add-on programs often have separate pricing.
Discount windows typically close 4–8 weeks before the event. They may coincide with hotel block cutoff dates, which can drive total trip cost more than the ticket price delta.
Historic patterns show that pre-conference workshops, certification-aligned trainings, and some networking events sell out first. Hotel inventory near the venue usually tightens next.
Some exhibitors also issue limited discount codes or bundled workshop passes.
Implication: Set a registration decision date aligned to your budget cycle. Secure hotel inventory early and track workshop capacity if advanced training is a priority.
The AI state of play: from generative to agentic — what matters for buyers
The 2025 shift is from “chat with your HR data” demos to agentic AI. Agents can plan, take actions across systems, and route exceptions to humans.
This increases impact but raises governance stakes around bias, safety, and auditability. The risks are highest in hiring and performance contexts.
Buyers should anchor evaluation to frameworks like the NIST AI Risk Management Framework. It emphasizes context-driven risks, measurement, and continuous monitoring.
In regulated contexts such as hiring, AI use must align with Title VII considerations. See the EEOC’s technical assistance on AI in hiring for details.
Implication: Make “agent design” and “control design” first-class requirements. Ensure every autonomous step is logged, reversible, and overseen by qualified humans.
Agentic AI evaluation checklist for HR teams
Agentic AI promises throughput and consistency, but only if governance keeps pace. Use this short checklist to standardize your reviews and de-risk pilots:
- Define the business outcome and guardrails: what decisions may the agent propose vs. auto-execute?
- Data minimization and lineage: document sources, retention, and training-data usage/opt-out policies.
- Fairness testing: pre-deployment and post-deployment bias tests on representative datasets; define acceptable thresholds.
- Human oversight: approval steps, exception queues, and role-based controls with time-stamped logs.
- Auditability: immutable logs for prompts, actions, model versions, and outcomes; exportable for auditors.
- Explainability: user-facing rationale for recommendations and access to model documentation/model cards.
- Safety and misuse prevention: instructions hardening, prompt filtering, and red-teaming results.
- Security posture: SOC 2 Type II and/or ISO/IEC 27001; SSO, SCIM, encryption in transit/at rest.
- Data residency and privacy: DPAs, Standard Contractual Clauses (SCCs) where applicable, PII handling, and deletion SLAs.
- Change management: versioning, approvals for model updates, rollback plans, and user enablement.
Implication: Make passing this checklist a gate for pilots and renewals. Weight fairness testing, auditability, and human-in-the-loop highest in hiring and performance workflows.
Security, privacy, and compliance signals to require
Your minimum bar should align to recognized security standards and transparent AI disclosures. Prioritize:
- Attestations: SOC 2 Type II (AICPA) and/or ISO/IEC 27001; request recent reports and mitigation plans. See AICPA SOC 2 overview.
- Privacy and data use: documented DPAs, subprocessor lists, data retention, and model training opt-out; encryption and key management.
- Access and identity: SSO (SAML/OIDC), SCIM provisioning, granular RBAC, field-level permissions, and IP allowlisting.
- AI disclosures: model sources, update cadence, safety approaches, and opt-outs for fine-tuning on your data.
- Regulatory alignment: EU AI Act readiness for “high-risk” employment use cases (see the EU AI Act overview) and Title VII-aligned selection testing (EEOC guidance linked above).
Implication: Bake these into your RFP and MSA as pass/fail requirements. Enforce remediation timelines and change-of-control protections.
Product launch roundup by category: TA, EX, payroll, analytics (GA vs beta vs roadmap)
The 2025 pattern is clear. More GA agentic features are shipping inside core platforms, with lighter-weight add-ons that price by usage.
For each announcement, capture status (GA vs beta vs roadmap) and pricing model (seat, usage, or bundled). Track dependencies (connectors, data permissions) and security attestations.
In practice, the fastest wins come from embedded assistants that shorten repetitive workflows. They avoid new data silos.
Conversely, beta-only features without admin controls or exportable logs often stall in enterprise reviews.
Implication: Track launch maturity and integration depth first. Defer betas that lack audit and admin controls to a sandbox environment.
Talent acquisition: sourcing, scheduling, assessments, and offers
TA launches are converging on end-to-end automation—from prospecting to interview scheduling to structured evaluations and calibrated offers.
Priorities include native integration to your ATS/CRM, LinkedIn Recruiter System Connect, calendar/email, and background screening.
Evaluate whether resume parsing, job matching, or scheduling agents are GA. Check what accuracy and throughput benchmarks are documented.
For assessments, insist on validity evidence and bias testing aligned to EEOC principles.
Implication: Pilot in high-volume requisitions with baseline metrics (time-to-slate, time-to-interview, candidate response rate). Aim to prove lift within 30–60 days.
Employee experience and performance: assistants, workflows, and feedback
EX/performance feature sets now bundle meeting summaries, goal/progress insights, feedback drafting, and nudges tied to performance cycles.
Look for integrations with collaboration suites, HRIS goals/OKRs, and learning catalogs. Admin controls for tone and data scope are essential.
Check if the assistant supports multilingual experiences, role-based personalization, and redaction for sensitive content.
For performance recommendations, require explainability and human-approval checkpoints.
Implication: Start with use cases that cut busywork without policy risk—status updates, recognition prompts, and review prep.
Payroll, compliance, and pay intelligence
In payroll and compliance, AI helps surface anomalies, forecast cash needs, and flag jurisdiction changes. Pay intelligence tools tackle pay equity and transparency obligations.
Ensure connectors to your HRIS, time systems, and GL. Verify that tax updates and compliance libraries are current.
For pay transparency, confirm job architecture alignment and market data sources. Tie guidance to your compensation philosophy.
Implication: Use AI-driven variance detection and equity monitoring as control layers. Set alert thresholds and escalation paths before go-live.
People analytics and skills intelligence
Analytics and skills-intelligence releases emphasize automated insights, skills inference, and planning across talent, learning, and workforce strategy.
The quality of outputs depends on identity resolution, data freshness, and a coherent skills ontology.
Ask vendors how skills are inferred, validated, and retired. Require exportable skills data and alignment options to existing taxonomies.
Implication: Make data prerequisites explicit. Define the minimal viable data pipeline (HRIS, ATS, LMS) before promising adoption targets.
Pitchfest 2025: finalists, judging criteria, winners, and enterprise readiness
Pitchfest showcases startups optimized for high-signal use cases. TA automation, EX co-pilots, and analytics/skills are perennial favorites.
While the stage rewards storytelling, enterprise readiness hinges on references, integrations, and security posture.
Expect judges to probe ROI math (time saved, quality improvements), durability (defensibility, margins), and risk (data, bias, security).
Implication: Use the event to accelerate diligence. Collect security docs, integration proofs, and two enterprise references before you shortlist.
Judging criteria and scoring methodology
Strong entries typically excel across five buckets: problem clarity, product differentiation, traction, defensibility, and trust.
Scoring often balances quantitative outcomes (cycle-time cuts, conversion lifts) with qualitative moats (unique data, hard-to-replicate workflows) and compliance posture.
Judges also weigh go-to-market efficiency and the ability to add value inside established suites rather than replace them. Expect direct questions on SOC 2/ISO status, model provenance, and how bias is measured and mitigated.
Implication: When you evaluate Pitchfest companies, mirror this rubric so your internal scorecards align to the judges’ standards.
Finalists and category mapping (with readiness signals)
Map each finalist to a buying category, then log readiness signals you can verify quickly:
- Category fit: TA automation, EX/performance, L&D/skills, analytics/compliance
- Readiness: SOC 2 Type II/ISO 27001; SSO/SCIM; DPA; subprocessor transparency
- Integrations: HRIS (Workday/Oracle/SAP), ATS (Greenhouse/Lever/Workday), collaboration (Microsoft/Google), data platforms
- References: two enterprise customers, deployment time, and realized KPIs
- Pricing: seat vs usage vs add-on; limits, overages, and renewal mechanics
Implication: Capture this on one page per startup. If two or more blanks persist (e.g., no SOC 2 and no references), relegate to watchlist instead of near-term POC.
Winners vs incumbents: capability and ROI contrasts
Winners often outperform incumbents on speed-to-value, configurability, or depth in a narrow workflow. Examples include interview scheduling or skills inference.
Incumbent suites win on breadth, native data access, and unified admin. They may lag in feature velocity or usability in specialized tasks.
Quantify tradeoffs in dollars and risk. A point solution that saves two recruiter hours per req with airtight ATS integration may beat a slower suite add-on, provided security and support match enterprise standards.
Implication: Use side-by-side ROI modeling that includes admin time, change costs, and integration debt—not just license fees.
Integrations and ecosystem announcements that matter
The most consequential news ties AI features to robust connectors and shared data models. Expect refreshed APIs, marketplace listings, and prebuilt skills graph connectors.
These reduce integration time and increase audit confidence.
Prioritize announcements that show event-driven integrations (webhooks), granular permissions, and identity reconciliation across HRIS, ATS, and collaboration tools.
Implication: Treat integration maturity as a gating factor. If logs aren’t exportable and identity isn’t reliable, AI value degrades fast.
Core HRIS suites and partner ecosystems
Look for expanded partner marketplaces, admin controls for AI modules, and open APIs that expose event streams (hire, transfer, comp change) securely.
Enterprise-grade connectors should support delta updates, retry logic, and lineage tracking. These features speed audits.
If a suite announces its own assistant, confirm scope boundaries (view-only vs act-on data) and approval gates. Check whether the assistant interacts with partners via APIs or embedded UI.
Implication: Prefer ecosystems that publish versioned APIs, rate-limit policies, and test sandboxes to stabilize downstream integrations.
ATS and CRM integrations
Momentum centers on native scheduling, messaging, and assessment flows. Recruiter co-pilots now draft outreach, summarize fit, and coordinate interviews.
The winners integrate deeply with calendars, email, and job boards while maintaining compliance logs.
Insist on LinkedIn Recruiter System Connect (RSC)-style integrations for recruiter tools. Require native referral and campaign tracking, and auditable evaluation forms synced to your ATS.
Implication: Measure success by cycle-time reductions and candidate experience scores. Verify nothing breaks your structured-interview discipline.
Data platforms, skills graphs, and identity
Expect more Snowflake/Databricks connectors, unified skills ontologies, and identity-resolution services. These reconcile person records across HRIS, ATS, LMS, and collaboration suites.
The payoff is better analytics and safer AI. That only holds if identity and permissions are consistent.
Demand clear mapping between vendor skills taxonomies and your internal frameworks. Require exportable graphs and API-level governance.
Implication: Make “identity and skills portability” a required line item. Avoid closed graphs that lock your data.
Funding and M&A around HR Tech 2025: deals and buyer implications
Capital flows cluster around automation that drives measurable productivity. Horizontal platforms with durable data moats also attract funding.
Private equity roll-ups continue in fragmented subcategories. Infrastructure costs are pushing vendors toward usage-based AI pricing.
Deal timing around conference week can signal roadmaps or runway extensions. Consolidation also raises integration and support risks post-acquisition.
Implication: Add a 30-day event-window watchlist for new funding and M&A. Revisit vendor risk scoring if ownership or burn profiles change.
Notable deals and strategic themes
Expect funding to favor agentic workflow layers such as TA scheduling and EX co-pilots. Data and skills platforms that improve multi-system analytics will also draw interest.
Strategic buyers will target adjacencies that round out suite gaps. Hot spots include assessments, pay intelligence, and onboarding.
Themes to watch include pricing experimentation for AI features and co-sell motions with hyperscalers or data platforms. Some startups will optimize for marketplace-led growth rather than direct procurement.
Implication: Vendor durability increasingly maps to distribution leverage and integration depth as much as novel features.
What buyers should watch (consolidation, runway, support)
When news hits, update your vendor scorecards with:
- Runway and support continuity: 18+ months preferred, or parent-company guarantees post-acquisition
- Product roadmap retention: commitments to existing APIs, SLAs, and feature parity timelines
- Data portability: export rights, schema docs, and practical exit plans
- Financial/legal protections: price-hold clauses, change-of-control termination rights, escrow for critical IP
Implication: Tie renewal terms to support continuity and data-export assurances. Bake in service credits for missed milestones during transitions.
Regulatory and compliance signals: EEOC AI hiring guidance, EU AI Act, and pay transparency
Three regulatory forces frame 2025 buying. Title VII risk in hiring algorithms, EU AI Act “high-risk” obligations, and expanding pay transparency rules.
The EEOC’s AI technical assistance underscores disparate impact risks and employer accountability. This holds even when vendors provide tools.
The EU AI Act overview signals conformity assessments, documentation, and oversight for high-risk employment systems serving EU workers.
Meanwhile, pay transparency initiatives, summarized by the U.S. Department of Labor’s pay transparency resources, elevate requirements for job postings, pay ranges, and fairness monitoring.
Implication: Align vendor selection with audit-readiness. Documentation, testing, and human oversight are nonnegotiable in hiring and performance contexts.
Risk controls and audit-readiness steps
Operationalize compliance with a durable control set:
- Inventory automated decision tools used in employment contexts and classify risk by use case and geography.
- Complete AI impact assessments (bias, explainability, data flows) before deployment; reassess on major model updates.
- Contract for fairness testing, documentation, and access to logs; align to NIST AI RMF functions (Map, Measure, Manage, Govern).
- Establish candidate/employee notices, consent where required, and clear human appeal mechanisms.
- Implement retention/deletion SLAs, subprocessor transparency, and data-subject request workflows.
- Create model change management: approvals, rollback plans, and stakeholder communication.
- Run quarterly monitors for drift and disparate impact; fix thresholds and escalation paths.
Implication: Make these controls shared OKRs across Legal, HR, and IT. Vendors that can’t support them should be excluded from high-risk use cases.
Buyer playbooks: 30-60-90 day plans for CHRO, TA, and HRIS leaders
Turn conference momentum into value by time-boxing decisions and pilots. The following playbooks align leadership accountability, metrics, and governance.
You can land wins without overextending risk.
Prioritize one to two agentic AI pilots per function with clear baselines and benefit hypotheses. Stage procurement to de-risk rollout.
Implication: Treat the 90-day window as a controlled experiment with executive visibility and go/no-go gates.
CHRO: operating model and governance-first adoption
- 30 days: Stand up an AI governance council (HR, Legal, InfoSec, Data) and ratify policies on acceptable use, human oversight, and bias testing. Approve an enterprise evaluation checklist and risk tiers.
- 60 days: Fund two pilots (one TA, one EX/performance) with measurable baselines; publish an ethics and transparency statement for employees and candidates.
- 90 days: Review pilot outcomes; scale one program with guardrails, and sunset or re-scope the other. Align 2025–2026 budgets to data infrastructure, identity, and skills portability.
KPIs: Policy adoption rate, pilot cycle-time reduction, fairness test pass rates, employee trust pulse (communications effectiveness).
TA leader: funnel friction, quality, and automation pilots
- 30 days: Baseline time-to-slate, time-to-interview, offer acceptance, and candidate NPS. Shortlist scheduling and outreach co-pilots with native ATS integrations.
- 60 days: Run a controlled pilot on 10–20 requisitions; enable structured interview kits and bias checks; require human approvals for agentic steps.
- 90 days: Expand to 50% of high-volume roles if cycle-time drops ≥20% with no fairness regressions; renegotiate contracts based on usage and realized outcomes.
KPIs: Time-to-interview, candidate response rate, interviewer utilization, pass/fail fairness checks, quality-of-hire proxies (first-90-day retention).
HRIS/People Ops: integrations, data quality, and TCO
- 30 days: Audit identity resolution across HRIS, ATS, and LMS; fix primary keys and SCIM provisioning. Create an integration error log and SLA.
- 60 days: Consolidate data flows to a governed warehouse; enable event-driven integrations and exportable AI logs. Publish admin standards for AI modules (permissions, data scope).
- 90 days: Decommission redundant connectors; enforce versioned APIs; finalize TCO model including AI usage, support, and change costs.
KPIs: Integration error rate, time-to-provision, data freshness, admin effort per rollout, cost per automated task.
ROI and procurement guardrails for new AI features
AI value hinges on matching pricing mechanics to usage patterns. You must also budget for change, data, and governance.
Most 2025 offers bundle AI as add-ons with usage ceilings. Total cost can exceed seat licenses once adoption scales.
Build a TCO model that includes enablement and ongoing fairness or safety monitoring. Do not focus on license fees alone.
Implication: Don’t greenlight “pilot pricing” without a path to sustainable unit economics at full adoption.
Pricing models, usage patterns, and hidden costs
Interrogate how you’ll pay and when costs spike:
- Per seat vs. usage: per-assistant or per-transaction; token-based models for generative features.
- Bundled add-ons: “Pro” tiers with monthly action caps; overage fees for interviews processed or messages sent.
- Hidden costs: data prep and integration, prompt engineering and testing, red-teaming, premium support, and data egress.
- Contract mechanics: seasonal scaling rights, price protection on model updates, and audit log access without extra fees.
Implication: Align model to real workflows. High-volume scheduling often favors usage bands; analytics assistants tied to analysts may suit seat pricing.
TCO and value realization checkpoints
Stage investment behind results:
- Gate 1 (Pilot complete): measurable lift on primary KPI (e.g., 20–30% cycle-time cut) and fairness/safety pass.
- Gate 2 (Limited rollout): admin overhead acceptable; logs export cleanly; integration errors below target.
- Gate 3 (Scale): unit economics at forecast; training complete; change requests within SLA.
Implication: Tie expansions and multi-year commitments to passing these gates. Include service credits for missed milestones.
Adoption and risk checklist for procurement
Standardize diligence so Legal, Security, and HR evaluate consistently:
- Security: SOC 2 Type II/ISO 27001, SSO/SCIM, encryption, vulnerability management
- Privacy: DPA, subprocessor transparency, data retention/deletion, training data opt-out
- AI governance: model documentation, bias testing, human-in-the-loop, audit logs
- Integration: supported connectors, event hooks, sandbox access, rate limits
- Commercials: usage definitions, overages, price-change protections, exit and portability rights
Implication: Make this checklist contractual. Attach it to the order form with explicit SLAs and remediation timelines.
How HR Tech 2025 compares to UNLEASH, Gartner ReimagineHR, and i4cp
HR Tech is the most vendor- and integration-dense of the four. It is ideal for hands-on demos and partner meetings.
UNLEASH balances thought leadership with a strong expo, especially in EMEA. Gartner ReimagineHR is research-led for senior leaders. i4cp centers on peer benchmarks and high-signal practices with limited expo noise.
Calendar-wise, HR Tech’s September timing makes it a decision catalyst for Q4 budgets. UNLEASH and Gartner often shape planning cycles and best-practice adoption.
Implication: Choose HR Tech for near-term buying and integration planning. Use Gartner/i4cp to refine strategy and operating models.
Who each event serves best (buyers, startups, investors)
- HR Tech: Buyers prioritizing near-term selection and integrations; startups seeking enterprise proof points; investors scanning product-market fit and category heat.
- UNLEASH: Global buyers and practitioners comparing regional vendors and approaches; strong EMEA networking.
- Gartner ReimagineHR: Senior leaders shaping strategy, operating models, and governance; limited hands-on technical depth.
- i4cp: People leaders focused on research-backed practices and peer exchange; minimal vendor noise.
Implication: Match event selection to your goals—transactions and integrations at HR Tech; strategic alignment and benchmarking at Gartner/i4cp; global scouting at UNLEASH.
Content depth, vendor mix, and timing
HR Tech offers the deepest vendor mix and the most robust integration talk track. UNLEASH brings international perspectives and innovative EX/skills stories.
Gartner concentrates on executive-grade frameworks. i4cp curates case studies with research rigor.
Timing differences let you iterate. Learn frameworks at Gartner/i4cp, shortlist at HR Tech, and validate alternatives at UNLEASH.
Implication: Plan a sequenced event strategy—frameworks first, buying second—so budgets and governance evolve with your stack.
Keynote takeaways and what leaders should do next
Keynotes consistently emphasize three imperatives: responsible AI, skills-based talent strategies, and measurable productivity.
The throughline is governance before growth. Build trust with transparency, then scale agentic workflows where auditability and fairness are provable.
For HR leaders, the next steps are clear. Codify AI policies, invest in identity and data quality, and focus pilots on high-friction workflows with clear KPIs.
Implication: Treat HR Tech 2025 as a 90-day catalyst. Convert insights into governed pilots, and tie 2026 budgets to the platforms and integrations that proved value under control.