GitLab Experiment Platform

A comprehensive experimentation platform for building data-driven organizations
GitLab Experiment is an enterprise-grade experimentation framework that enables teams to validate hypotheses, optimize user experiences, and make evidence-based product decisions at scale. Built on years of production experience at GitLab, this platform provides the foundation for a mature experimentation culture across your entire organization.
At GitLab, we run experiments as A/B/n tests and review the data they generate. From that data, we determine the best performing code path and promote it as the new default, or revert back to the original code path. You can read our Experiment Guide to learn how we use this gem internally at GitLab.
[[TOC]]
Why GitLab Experiment?
Built for Scale and Reliability
- Production-tested at GitLab scale with millions of users
- Type-safe and testable with comprehensive RSpec support
- Framework agnostic - works with Rails, Sinatra, or standalone Ruby applications
- Redis-backed caching for consistent user experiences across sessions
- GDPR-compliant with anonymous tracking and built-in DNT (Do Not Track) support
Designed for Teams
- Developer-friendly DSL that reads like natural language
- Organized experiment classes that live alongside your application code
- Built-in generators for rapid experiment creation
- Comprehensive testing support with custom RSpec matchers
- Rails integration with automatic middleware mounting and view helpers
Enterprise-Ready Features
- Flexible rollout strategies (percent-based, random, round-robin, or custom)
- Advanced segmentation to target specific user populations
- Multi-variant testing (A/B/n) with unlimited experimental paths
- Progressive rollouts with the
only_assignedfeature - Context migrations for evolving experiments without losing data
- Integration-ready with existing feature flag systems (Flipper, Unleash, etc.)
Use Cases Across Your Organization
Product Teams: Optimize User Experiences
- Onboarding flows: Test different signup sequences to maximize activation
- UI/UX changes: Validate design decisions with real user behavior data
- Feature rollouts: Gradually release features to measure impact before full deployment
- Pricing experiments: Test different pricing strategies and messaging
Growth Teams: Drive Conversion
- Call-to-action optimization: Test button colors, copy, and placement
- Landing page variations: Experiment with different value propositions
- Email campaigns: A/B test subject lines and content
- Trial conversion: Optimize paths from trial to paid subscriptions
Engineering Teams: Safe Deployments
- Performance optimizations: Compare algorithm implementations under real load
- Architecture changes: Validate new code paths before full migration
- API versions: Run multiple API implementations side-by-side
- Infrastructure experiments: Test different caching or database strategies
Data Science Teams: Recommendation Systems
- Algorithm tuning: Compare ML model variations in production
- Personalization: Test different recommendation strategies
- Search ranking: Optimize search results based on user engagement
- Content discovery: Experiment with different content surfaces
Platform Capabilities
Core Experimentation Features
Multi-variant Testing (A/B/n) Run experiments with any number of variants, not just A/B tests. Perfect for testing multiple approaches simultaneously.
Smart Segmentation Route specific user populations to predetermined variants based on business rules, ensuring consistent experiences for targeted groups.
Progressive Rollouts
Use the only_assigned feature to show experimental features only to users already in the experiment, enabling
controlled expansion.
Context Flexibility Experiments can be sticky to users, projects, organizations, or any combination - enabling complex scenarios beyond user-centric testing.
Anonymous Tracking Built-in privacy protection with anonymous context keys, automatic cookie migration, and GDPR compliance.
Automatic Assignment Tracking
Every experiment automatically tracks an :assignment event when it runs - zero configuration required. Combined with
the anonymous context key, this gives your data team a complete picture of variant distribution and funnel entry without
any additional instrumentation.
Client-Side Integration Seamlessly extend experiments to the frontend with JavaScript integration, enabling full-stack experimentation.
Inline and Class-Based APIs Define experiments inline with blocks for quick iterations, or use dedicated experiment classes for complex logic - or combine both. Class-based experiments define default behaviors that can be overridden inline at any call site, giving teams the flexibility to start simple and evolve without rewriting:
experiment(:pill_color, actor: current_user) do |e|
e.control { '<strong>control</strong>' }
end
Context as a Design Framework Context is the most important design decision in any experiment. It determines stickiness, cache behavior, and how events are correlated. Choose per-user context for personalization experiments, per-project for infrastructure tests, per-group for organizational rollouts, or combine dimensions for precision targeting. This flexibility enables experimentation strategies that go far beyond simple user-centric A/B tests.
Decoupled Assignment with Publish
Surface experiment assignments to the client layer without executing server-side behavior using publish. This enables
frontend-only experiments, pre-assignment in before_action hooks, and scenarios where variant data needs to be
available across the stack without triggering server-side code paths:
before_action -> { experiment(:pill_color, actor: current_user).publish }, only: [:show]
Integration Ecosystem
Feature Flag Integration Connect with existing feature flag systems like Flipper or Unleash through custom rollout strategies.
Analytics Integration Flexible tracking callbacks integrate with any analytics platform - Snowplow, Amplitude, Mixpanel, or your data warehouse.
Monitoring and Observability Built-in logging and callbacks for integration with APM tools and monitoring systems.
Email and Markdown Special middleware for tracking experiments in email links and static content.
Terminology
When we discuss the platform, we use specific terms that are worth understanding:
- experiment - Any deviation of code paths we want to test
- context - Identifies a consistent experience (user, project, session, etc.)
- control - The default or "original" code path
- candidate - One experimental code path (used in A/B tests)
- variant(s) - Multiple experimental paths (used in A/B/n tests)
- behaviors - All possible code paths (control + all variants)
- rollout strategy - Logic determining if an experiment is enabled and how variants are assigned
- segmentation - Rules for routing specific contexts to predetermined variants
- exclusion - Rules for keeping contexts out of experiments entirely
Quick Start: From Zero to Experiment in 5 Minutes
Installation
Add the gem to your Gemfile and then bundle install.
gem 'gitlab-experiment'
If you're using Rails, install the initializer which provides basic configuration, documentation, and the base experiment class:
$ rails generate gitlab:experiment:install
Your First Experiment
Let's create a real-world experiment to optimize a call-to-action button. This example demonstrates the power of the platform while remaining practical.
Step 1: Generate the experiment
Hypothesis: A more prominent call-to-action button will increase conversion rates
$ rails generate gitlab:experiment signup_cta
This creates app/experiments/signup_cta_experiment.rb with helpful inline documentation.
Step 2: Define your experiment class
class SignupCtaExperiment < ApplicationExperiment
# Define the control (current experience)
control { 'btn-default' }
# Define the candidate (new experience to test)
candidate { 'btn-primary btn-lg' }
# Optional: Exclude certain users
exclude :existing_customers
# Optional: Track when the experiment runs
after_run :log_experiment_assignment
private
def existing_customers
context.actor&.subscribed?
end
def log_experiment_assignment
Rails.logger.info("User assigned to #{assigned.name} variant")
end
end
Step 3: Use the experiment in your view
-# The experiment is sticky to the current user
-# Anonymous users get a cookie-based assignment
%button{ class: experiment(:signup_cta, actor: current_user).run }
Start Free Trial
Step 4: Track engagement
# In your controller, track when users click the button
def create_trial
experiment(:signup_cta, actor: current_user).track(:signup_completed)
# ... rest of your trial creation logic
end
That's it! Your experiment is now running, collecting data, and providing consistent experiences to your users.
Real-World Examples
Example 1: Onboarding Flow Optimization
Business Context: Product team wants to increase new user activation by testing different onboarding sequences.
class OnboardingFlowExperiment < ApplicationExperiment
# Three different onboarding approaches
control { :standard_tour } # Current 5-step tour
variant(:quick) { :quick_start } # Streamlined 2-step flow
variant(:video) { :video_guide } # Video-based walkthrough
# Only show to new users who haven't completed onboarding
exclude :has_completed_onboarding
# Segment enterprise trial users to the standard tour
segment :enterprise_trial?, variant: :control
private
def has_completed_onboarding
context.actor&.onboarding_completed_at.present?
end
def enterprise_trial?
context.actor&.trial_type == 'enterprise'
end
end
# In your onboarding controller
def show
flow = experiment(:onboarding_flow, actor: current_user).run
render_onboarding_flow(flow)
end
# Track completion
def complete
experiment(:onboarding_flow, actor: current_user).track(:completed)
# ... mark user as onboarded
end
Example 2: Pricing Page Experiment
Business Context: Growth team wants to test whether showing annual savings increases annual plan selection.
class PricingDisplayExperiment < ApplicationExperiment
control { :monthly_default }
candidate { :annual_default_with_savings }
# Only run for unauthenticated visitors
exclude :authenticated_user
private
def authenticated_user
context.actor.present?
end
end
# In your pricing view
- pricing_variant = experiment(:pricing_display, actor: current_user).run
= render "pricing/#{pricing_variant}"
# Track plan selections
def select_plan
experiment(:pricing_display, actor: current_user).track(:plan_selected,
value: params[:plan_type] == 'annual' ? 1 : 0
)
end
Example 3: Algorithm Performance Test
Business Context: Engineering team wants to compare a new search algorithm's performance before full rollout.
class SearchAlgorithmExperiment < ApplicationExperiment
control { SearchEngine::Legacy }
candidate { SearchEngine::Neural }
# Only run for 25% of searches
default_rollout :percent, distribution: { control: 75, candidate: 25 }
# Exclude searches from API (higher SLA requirements)
exclude :api_request
# Track performance metrics
after_run :record_search_timing
private
def api_request
context.request&.path&.start_with?('/api/')
end
def record_search_timing
# Custom metrics tracking
end
end
# In your search service
def search(query)
algorithm = experiment(:search_algorithm,
actor: current_user,
project: current_project
).run
results = algorithm.search(query)
experiment(:search_algorithm,
actor: current_user,
project: current_project
).track(:search_completed, value: results.count)
results
end
Example 4: Progressive Feature Rollout
Business Context: Launching a new AI-assisted code review feature, want to expand gradually to manage load and gather feedback.
class AiCodeReviewExperiment < ApplicationExperiment
control { false } # Feature disabled
candidate { true } # Feature enabled
# Start with 5% rollout
default_rollout :percent, distribution: { control: 95, candidate: 5 }
# Segment beta program users to always get the feature
segment :beta_user?, variant: :candidate
# Exclude free tier (computational cost consideration)
exclude :free_tier_user
private
def beta_user?
context.actor&.beta_features_enabled?
end
def free_tier_user
context.actor&.subscription_tier == 'free'
end
end
# In your merge request view
- if experiment(:ai_code_review, actor: current_user, project: @project).run
.ai-code-review-panel
= render 'ai_suggestions'
# Track usage
def apply_ai_suggestion
experiment(:ai_code_review, actor: current_user, project: @project)
.track(:suggestion_applied)
end
Platform Integration Patterns
Integration with Feature Flags (Flipper)
Many organizations already use feature flag systems. GitLab Experiment integrates seamlessly:
module Gitlab::Experiment::Rollout
class Flipper < Percent
def enabled?
::Flipper.enabled?(experiment.name, experiment_actor)
end
def experiment_actor
Struct.new(:flipper_id).new("Experiment;#{id}")
end
end
end
# Configure globally
Gitlab::Experiment.configure do |config|
config.default_rollout = Gitlab::Experiment::Rollout::Flipper.new
end
# Now Flipper controls your experiments
Flipper.enable_percentage_of_actors(:signup_cta, 50)
Integration with Analytics Platforms
Connect experiments to your analytics stack:
Gitlab::Experiment.configure do |config|
config.tracking_behavior = lambda do |event_name, **data|
# Snowplow
SnowplowTracker.track_struct_event(
category: 'experiment',
action: event_name,
property: data[:experiment],
context: [{ schema: 'experiment_context', data: data }]
)
# Amplitude (example)
Amplitude.track(
user_id: data[:key], # Anonymous experiment key
event_type: "experiment_#{event_name}",
event_properties: data
)
# Custom data warehouse
DataWarehouse.log_experiment_event(event_name, data)
end
end
Multi-Application Consistency
Share experiment assignments across multiple applications:
# Shared Redis cache
Gitlab::Experiment.configure do |config|
config.cache = Gitlab::Experiment::Cache::RedisHashStore.new(
Redis.new(url: ENV['REDIS_URL']),
expires_in: 30.days
)
end
# Now experiments stay consistent across your web app, API, and background jobs
Advanced Features
Multi-Variant (A/B/n) Testing
Test multiple variations simultaneously to find the optimal solution:
class NotificationStyleExperiment < ApplicationExperiment
# Test three different notification approaches
control { :banner } # Current: banner at top
variant(:toast) { :toast } # Toast notification
variant(:modal) { :modal } # Modal dialog
# Distribute traffic evenly across all three
default_rollout :percent,
distribution: { control: 34, toast: 33, modal: 33 }
# Exclude mobile users (different UI constraints)
exclude :mobile_user
# Segment power users to toast (less intrusive)
segment :power_user?, variant: :toast
private
def mobile_user
context.request&.user_agent&.match?(/Mobile/)
end
def power_user?
context.actor&.actions_count > 1000
end
end
Exclusion Rules
Keep contexts out of experiments entirely based on business rules:
class FeatureExperiment < ApplicationExperiment
# Exclude existing customers (only test on prospects)
exclude :existing_customer
# Exclude during maintenance windows
exclude -> { context.project&.under_maintenance? }
# Exclude if feature is explicitly disabled
exclude :feature_disabled
private
def existing_customer
context.actor&.subscribed?
end
def feature_disabled
!FeatureFlag.enabled?(:allow_experiment, context.actor)
end
end
Key behaviors:
- Excluded users always receive the control experience
- No tracking events are recorded for excluded users
- Exclusion rules are evaluated in order, first match wins
- Exclusions improve performance by exiting early
Inline exclusion is also supported:
experiment(:feature, actor: current_user) do |e|
e.exclude! unless can?(current_user, :manage, project)
e.control { 'standard' }
e.candidate { 'enhanced' }
end
Note: Although tracking calls will be ignored on all exclusions, you may want to check exclusion yourself in expensive
custom logic by calling the should_track? or excluded? methods.
Note: When using exclusion rules it's important to understand that the control assignment is cached, which improves future experiment run performance but can be a gotcha around caching.
Note: Exclusion rules aren't the best way to determine if an experiment is enabled. There's an enabled? method that
can be overridden to have a high-level way of determining if an experiment should be running and tracking at all. This
enabled? check should be as efficient as possible because it's the first early opt out path an experiment can
implement. This can be seen in How Experiments Work (Technical).
Segmentation Rules
Route specific populations to predetermined variants:
class NewFeatureExperiment < ApplicationExperiment
# Route VIP customers to the new feature
segment :vip_customer?, variant: :candidate
# Route enterprise trial users to the enhanced experience
segment :enterprise_trial?, variant: :candidate
# Route users from specific campaigns to specific variants
segment(variant: :candidate) { context.campaign == 'product_launch_2024' }
private
def vip_customer?
context.actor&.account_value > 100_000
end
def enterprise_trial?
context.actor&.trial_tier == 'enterprise'
end
end
Key behaviors:
- Segmentation rules are evaluated in order, first match wins
- Segmented assignments are cached for consistency
- Perfect for gradually expanding successful experiments
- Enables sophisticated population targeting
Lifecycle Callbacks
Execute custom logic at different stages of experiment execution:
class PerformanceExperiment < ApplicationExperiment
# Run before the variant is determined
before_run :log_experiment_start
# Run after the variant is executed
after_run :record_timing_metrics, :notify_analytics_team
# Wrap the entire execution
around_run do |experiment, block|
start_time = Time.current
result = block.call
duration = Time.current - start_time
Metrics.record("experiment.#{experiment.name}.duration", duration)
result
end
private
def log_experiment_start
Rails.logger.info("Starting experiment: #{name}")
end
def record_timing_metrics
# Custom timing logic
end
def notify_analytics_team
# Send to analytics platform
end
end
Use cases for callbacks:
- Performance monitoring and APM integration
- Custom analytics and data warehouse updates
- Experiment-specific logging and debugging
- Integration with external systems
Progressive Rollout with only_assigned
Control experiment expansion by only showing features to users already assigned to the experiment. This is critical for managing blast radius and controlled rollouts:
The Challenge: You launch an experiment to 10% of new signups. Later, you want to show experimental features on other pages, but only to users already in the experiment - not expand to 10% of all users across the platform.
The Solution: Use only_assigned: true
# Step 1: Assign users during signup (10% of new signups)
class RegistrationsController < ApplicationController
def create
user = User.create!(user_params)
# This assigns 10% to candidate, 90% to control
experiment(:onboarding_v2, actor: user).publish
redirect_to dashboard_path
end
end
# Step 2: Later, show features only to those already assigned
class DashboardController < ApplicationController
def show
# This will NOT expand the experiment to 10% of all users
# Only users assigned in Step 1 will see the experimental UI
@show_new_features = experiment(:onboarding_v2,
actor: current_user,
only_assigned: true
).assigned.name == 'candidate'
end
end
# Step 3: Show UI conditionally across the app
- if experiment(:onboarding_v2, actor: current_user, only_assigned: true).run
.new-onboarding-features
= render 'enhanced_dashboard'
Behavior with only_assigned: true:
- ✅ If user already assigned → returns their cached variant
- ✅ If user not assigned → returns control, no tracking
- ✅ Experiment reach stays controlled
- ✅ Perfect for multi-page experimental experiences
Real-world use cases:
- Post-signup experiences: Assign at signup, show features throughout the app
- Gradual feature expansion: Roll out to 5%, then add more touchpoints without expanding population
- Cleanup phases: Maintain experience for existing participants while preventing new assignments
- A/B testing with multiple surfaces: Test a hypothesis across multiple pages without assignment leakage
Custom Rollout Strategies
The platform supports multiple rollout strategies out of the box, and you can create custom strategies for your specific needs.
Built-in strategies:
Percent- Consistent percentage-based assignment (default, recommended)Random- True random assignment (useful for load testing)RoundRobin- Cycle through variants (requires caching)Base- Useful for building custom rollout strategies
class LoadTestExperiment < ApplicationExperiment
# Randomly test two different caching strategies
default_rollout :random
control { CacheStrategy::Redis }
candidate { CacheStrategy::Memcached }
end
class GradualRolloutExperiment < ApplicationExperiment
# Start with 5% in the new experience
default_rollout :percent,
distribution: { control: 95, candidate: 5 }
end
See the Advanced: Custom Rollout Strategies section for building your own integration with feature flag systems.
Organizational Best Practices
Experiment Lifecycle Management
1. Hypothesis Formation
# Document your hypothesis in the experiment class
class CheckoutFlowExperiment < ApplicationExperiment
# Hypothesis: Reducing checkout steps from 3 to 2 will increase completion rate
# Success metric: 5% increase in checkout completion
# Target: All free trial users
# Duration: 2 weeks
# Owner: @growth-team
control { :three_step_checkout }
candidate { :two_step_checkout }
exclude :existing_customer
end
2. Gradual Rollout
# Week 1: 5% rollout
default_rollout :percent, distribution: { control: 95, candidate: 5 }
# Week 2: Increase to 25% if metrics look good
default_rollout :percent, distribution: { control: 75, candidate: 25 }
# Week 3: Full rollout if successful
default_rollout :percent, distribution: { control: 0, candidate: 100 }
3. Monitoring and Alerting
class CriticalPathExperiment < ApplicationExperiment
after_run :monitor_performance
after_run :alert_on_errors
private
def monitor_performance
Metrics.increment("experiment.#{name}.#{assigned.name}")
end
def alert_on_errors
if context.error_rate > threshold
PagerDuty.alert("High error rate in #{name}")
end
end
end
4. Experiment Cleanup
# When experiment is conclusive, clean up:
# 1. Remove the experiment code
# 2. Promote winner to production
# 3. Document learnings
# Before cleanup, archive results:
experiment(:checkout_flow).publish
# Export data for historical analysis
Team Collaboration Patterns
Product + Engineering + Data Science
class CollaborativeExperiment < ApplicationExperiment
# Product defines the hypothesis and variants
control { :current_flow }
candidate { :new_flow }
# Engineering defines segmentation and rollout
segment :beta_users, variant: :candidate
default_rollout :percent, distribution: { control: 90, candidate: 10 }
# Data science defines tracking and metrics
after_run :track_funnel_step
def track_funnel_step
Analytics.track_experiment_step(
experiment: name,
variant: assigned.name,
funnel_position: context.step,
user_segment: context.actor&.segment
)
end
end
Testing Strategy
Write tests for your experiments using the included RSpec matchers:
RSpec.describe CheckoutFlowExperiment do
describe 'segmentation' do
it 'routes existing customers to control' do
customer = create(:user, :with_subscription)
expect(experiment(:checkout_flow)).to exclude(actor: customer)
end
it 'routes enterprise trials to candidate' do
trial = create(:user, :enterprise_trial)
expect(experiment(:checkout_flow))
.to segment(actor: trial).into(:candidate)
end
end
describe 'tracking' do
it 'tracks checkout completion' do
expect(experiment(:checkout_flow)).to track(:completed)
.on_next_instance
CheckoutService.complete(user: user)
end
end
end
Naming Conventions
Establish clear naming conventions for your organization:
# Good: Descriptive experiment names
class OnboardingFlowV2Experiment < ApplicationExperiment; end
class PricingPageAnnualFocusExperiment < ApplicationExperiment; end
class SearchAlgorithmNeuralExperiment < ApplicationExperiment; end
# Avoid: Vague names
class TestExperiment < ApplicationExperiment; end # What are we testing?
class ExperimentOne < ApplicationExperiment; end # No context
Technical Reference
How Experiments Work (Technical)
Understanding the experiment resolution flow helps you design better experiments and debug issues:
Decision tree for variant assignment:
graph TD
GP[General Pool/Population] --> Running?[Rollout Enabled?]
Running? -->|Yes| Forced?[Forced Assignment?]
Running? -->|No| Excluded[Control / No Tracking]
Forced? -->|Yes / Cached| ForcedVariant[Forced Variant]
Forced? -->|No| Cached?[Cached? / Pre-segmented?]
Cached? -->|No| Excluded?
Cached? -->|Yes| Cached[Cached Value]
Excluded? -->|Yes / Cached| Excluded
Excluded? -->|No| Segmented?
Segmented? -->|Yes / Cached| VariantA
Segmented? -->|No| Rollout[Rollout Resolve]
Rollout --> Control
Rollout -->|Cached| VariantA
Rollout -->|Cached| VariantB
Rollout -->|Cached| VariantN
class ForcedVariant,VariantA,VariantB,VariantN included
class Control,Excluded excluded
class Cached cached
Key points:
- Rollout must be enabled for any variant assignment (including forced assignment)
- Forced assignment takes priority over cache/exclusion/segmentation (via
glex_forcequery parameter) - Cache provides consistency across calls
- Segmentation takes priority over rollout
only_assigned: trueexits early if no cache hit
Experiment Context and Stickiness
Internally, experiments have what's referred to as the context "key" that represents the unique and anonymous id of a given context. This allows us to assign the same variant between different calls to the experiment, is used in caching and can be used in event data downstream. This context "key" is how an experiment remains "sticky" to a given context, and is an important aspect to understand.
Context defines stickiness - experiments remain consistent by generating an anonymous key from the context:
# Sticky to user - same user gets same variant everywhere
experiment(:feature, actor: current_user)
# Sticky to project - all users on a project get the same experience
experiment(:feature, project: project)
# Sticky to user+project - same user gets same variant per project
experiment(:feature, actor: current_user, project: project)
# Custom stickiness - explicitly define what creates consistency
experiment(:feature, actor: current_user, project: project, sticky_to: project)
The actor keyword has special behavior:
- Anonymous users → temporary cookie-based assignment
- Upon sign-in → cookie migrates to user ID automatically
- Enables consistent experience across anonymous → authenticated journey
Using Experiments Beyond Views
By default, Gitlab::Experiment injects itself into the controller, view, and mailer layers. This exposes the
experiment method application wide in those layers. Some experiments may extend outside of those layers however, so
you may want to include it elsewhere. For instance in an irb session or the rails console, or in all your service
objects, background jobs, or similar:
# In all background jobs
class ApplicationJob < ActiveJob::Base
include Gitlab::Experiment::Dsl
end
# In service objects
class ApplicationService
include Gitlab::Experiment::Dsl
end
# In a console session
include Gitlab::Experiment::Dsl
experiment(:feature, actor: User.first).run
Manual Variant Assignment
You can also specify the variant manually...
Generally, defining segmentation rules is a better way to approach routing into specific variants, but it's possible to explicitly specify the variant when running an experiment. Caching: It's important to understand what this might do to your data during rollout, so use this with careful consideration. Any time a specific variant is assigned manually, or through segmentation (including `:control`) it will be cached for that context. That means that if you manually assign `:control`, that context will never be moved out of the control unless you do it programmatically elsewhere. ```ruby include Gitlab::Experiment::Dsl # Assign the candidate manually. ex = experiment(:pill_color, :red, actor: User.first) # => #Forced Variant Assignment (QA/UAT)
For testing and validation purposes, you can force a specific variant assignment via a URL query parameter. This is useful for QA testing in staging or production environments where you need to verify a specific variant's behavior.
Configuration:
Forced assignment is disabled by default. Enable it in your initializer:
Gitlab::Experiment.configure do |config|
config.allow_forced_assignment = true
end
Usage:
Append the glex_force query parameter to any URL with the format experiment_name:variant_name:
https://your-app.com/signup?glex_force=myapp_signup_cta:candidate
The forced variant is written to the cache (Redis) on the same request, making it permanent for that context. The query parameter only needs to be provided once -- after that, the variant is persisted in the cache like any normal assignment.
Anonymous user (nil actor) -- initial assignment
This is the primary use case for QA testing signup flows and landing pages. The user is not signed in, so the actor is nil and the experiment uses a cookie-based context key.
- Anonymous user visits
https://your-app.com/signup?glex_force=signup_cta:candidate - The forced variant
:candidateis written to Redis under the cookie-based context key - The user signs in -- the standard cookie migration carries the forced variant to their real identity
- All future requests use
:candidatefrom Redis, permanently
This means a QA tester can force a variant before signup and have it follow the user through the entire anonymous-to-authenticated journey.
Signed-in user -- initial assignment
When a signed-in user hasn't been assigned a variant yet, the force param assigns and caches it immediately:
https://your-app.com/dashboard?glex_force=new_feature:candidate
The variant is cached under the user's context key on this request. No further query parameter is needed.
Signed-in user -- re-assignment (overwriting an existing variant)
If a user was previously assigned :control (by the rollout strategy or a prior force), the force param overwrites
the cached value:
https://your-app.com/dashboard?glex_force=new_feature:candidate
The existing :control assignment in Redis is replaced with :candidate. This is useful when QA needs to switch a
user between variants without clearing the cache manually.
Disabled experiments and feature flags
Forced assignment requires the experiment to be enabled. If the experiment is disabled (as determined by the rollout
strategy's enabled? method), the glex_force parameter is ignored and normal resolution applies (which will assign
control).
This is intentional -- a disabled experiment may be disabled for valid reasons (incomplete implementation, known issues, compliance constraints, etc.) and force assignment should not provide a way to bypass that decision. To use forced assignment, ensure the experiment is enabled first through your rollout strategy.
Important notes:
- The experiment name in the parameter must match the full experiment name (including any configured
name_prefix). - If the variant name doesn't match a registered behavior, the forced assignment is ignored and normal variant resolution proceeds (typically resulting in the control variant).
- Forced assignment does not override a variant that was already set via the constructor or an explicit
assigned()call within the same request. - This feature requires a
requestobject withparamsto be available in the experiment context.
[!NOTE] Because forcing the variant ignores the exclusion/segmentation process it will cover up those types of errors so if your experiment relies on these types of logic this testing method should be avoided.
Experiment Signature
The best way to understand the details of an experiment is through its signature. An example signature can be retrieved
by calling the signature method, and looks like the following:
experiment(:example).signature # => {:variant=>"control", :experiment=>"example", :key=>"4d7aee..."}
An experiment signature is useful when tracking events and when using experiments on the client layer. The signature can
also contain the optional migration_keys, and excluded properties.
Return Value
By default the return value of calling experiment is a Gitlab::Experiment instance, or whatever class the
experiment is resolved to, which likely inherits from Gitlab::Experiment. In simple cases you may want only the
results of running the experiment though. You can call run within the block to get the return value of the assigned
variant.
# Normally an experiment instance.
experiment(:example) do |e|
e.control { 'A' }
e.candidate { 'B' }
end # => #<Gitlab::Experiment:0x...>
# But calling `run` causes the return value to be the result.
experiment(:example) do |e|
e.control { 'A' }
e.candidate { 'B' }
e.run
end # => 'A'
Context migrations
There are times when we need to change context while an experiment is running. We make this possible by passing the migration data to the experiment.
Take for instance, that you might be using version: 1 in your context currently.
To migrate this to version: 2, provide the portion of the context you wish to change using a migrated_with option.
In providing the context migration data, we can resolve an experience and its events all the way back. This can also help in keeping our cache relevant.
# First implementation.
experiment(:example, actor: current_user, version: 1)
# Migrate just the `:version` portion.
experiment(:example, actor: current_user, version: 2, migrated_with: { version: 1 })
You can add or remove context by providing a migrated_from option.
This approach expects a full context replacement -- i.e. what it was before you added or removed the new context key.
If you wanted to introduce a version to your context, provide the full previous context.
# First implementation.
experiment(:example, actor: current_user)
# Migrate the full context of `{ actor: current_user }` to `{ actor: current_user, version: 1 }`.
experiment(:example, actor: current_user, version: 1, migrated_from: { actor: current_user })
When you migrate context, this information is included in the signature of the experiment. This can be used downstream in event handling and reporting to resolve a series of events back to a single experience, while also keeping everything anonymous.
An example of our experiment signature when we migrate would include the migration_keys property:
ex = experiment(:example, version: 1)
ex.signature # => {:key=>"20d69a...", ...}
ex = experiment(:example, version: 2, migrated_from: { version: 1 })
ex.signature # => {:key=>"9e9d93...", :migration_keys=>["20d69a..."], ...}
Cookies and the actor keyword
We use cookies to auto migrate an unknown value into a known value, often in the case of the current user.
The implementation of this uses the same concept outlined above with context migrations, but will happen automatically
for you if you use the actor keyword.
When you use the actor: current_user pattern in your context, the nil case is handled by setting a special cookie for
the experiment and then deleting the cookie, and migrating the context key to the one generated from the user when
they've signed in.
This cookie is a temporary, randomized uuid and isn't associated with a user. When we can finally provide an actor, the context is auto migrated from the cookie to that actor.
# The actor key is not present, so no cookie is set.
experiment(:example, project: project)
# The actor key is present but nil, so the cookie is set and used.
experiment(:example, actor: nil, project: project)
# The actor key is present and isn't nil, so the cookie value (if found) is
# migrated forward and the cookie is deleted.
experiment(:example, actor: current_user, project: project)
Note: The cookie is deleted when resolved, but can be assigned again if the actor is ever nil again.
A good example of this scenario would be on a sign in page.
When a potential user arrives, they would never be known, so a cookie would be set for them, and then resolved/removed
as soon as they signed in.
This process would repeat each time they arrived while not being signed in and can complicate reporting unless it's
handled well in the data layers.
Note: To read and write cookies, we provide the request from within the controller and views.
The cookie migration will happen automatically if the experiment is within those layers.
You'll need to provide the request as an option to the experiment if it's outside of the controller and views.
experiment(:example, actor: current_user, request: request)
Note: For edge cases, you can pass the cookie through by assigning it yourself -- e.g. actor:
request.cookie_jar.signed['example_id'].
The cookie name is the full experiment name (including any configured prefix) with _id appended -- e.g.
pill_color_id for the PillColorExperiment.
Client layer
Experiments that have been run (or published) during the request lifecycle can be pushed into to the client layer by injecting the published experiments into javascript in a layout or view using something like:
= javascript_tag(nonce: content_security_policy_nonce) do
window.experiments = #{raw ApplicationExperiment.published_experiments.to_json};
The window.experiments object can then be used in your client implementation to determine experimental behavior at
that layer as well.
For instance, we can now access the window.experiments.pill_color object to get the variant that was assigned, if the
context was excluded, and to use the context key in our client side events.
Adoption Guide for Organizations
Phase 1: Foundation (Week 1-2)
- Install and configure the gem
- Set up analytics integration in the initializer
- Create a base experiment class for your organization
- Run your first small experiment (low-risk, high-visibility)
Phase 2: Team Enablement (Week 3-4)
- Document your organization's patterns (naming, testing, rollout)
- Train teams on experiment lifecycle
- Establish experiment review process (hypothesis → implementation → analysis)
- Run 2-3 experiments across different teams
Phase 3: Scale (Month 2+)
- Integrate with feature flag system (if applicable)
- Build dashboards for experiment monitoring
- Establish data review cadence (weekly experiment reviews)
- Scale to 5-10 concurrent experiments
Common Pitfalls to Avoid
❌ Don't: Run experiments without clear success metrics
class VagueExperiment < ApplicationExperiment
# What are we trying to learn?
control { :old_way }
candidate { :new_way }
end
✅ Do: Document hypothesis and success criteria
class CheckoutOptimizationExperiment < ApplicationExperiment
# Hypothesis: Showing trust badges increases checkout completion
# Success Metric: 5% increase in completion rate
# Target: Free trial users
# Duration: 2 weeks
control { :without_badges }
candidate { :with_trust_badges }
end
❌ Don't: Let experiments run indefinitely
- Set time bounds for every experiment
- Review results at planned intervals
- Make a decision: promote winner, revert, or iterate
✅ Do: Build experiment cleanup into your process
- Schedule experiment review meetings
- Archive experiment results
- Clean up experiment code after conclusion
Platform Configuration
The platform requires initial configuration to integrate with your analytics and infrastructure.
Basic configuration (in config/initializers/gitlab_experiment.rb):
Gitlab::Experiment.configure do |config|
# How experiment events are tracked
config.tracking_behavior = lambda do |event_name, **data|
YourAnalytics.track(
user_id: data[:key], # Anonymous experiment key
event: "experiment_#{event_name}",
properties: data
)
end
# How experiments are cached (recommended: Redis)
config.cache = Gitlab::Experiment::Cache::RedisHashStore.new(
Redis.new(url: ENV['REDIS_URL']),
expires_in: 30.days
)
# Optional: Prefix all experiment names
config.name_prefix = 'mycompany'
# Optional: Default rollout strategy
config.default_rollout = Gitlab::Experiment::Rollout::Percent.new
end
See the complete initializer template for all configuration options.
Advanced: Caching Configuration
Why caching matters:
- Ensures consistent user experience across sessions
- Improves performance (skip rollout logic after first assignment)
- Required for
only_assignedfunctionality - Enables context migrations
Cache options:
# Option 1: Use Rails cache (simple)
Gitlab::Experiment.configure do |config|
config.cache = Rails.cache
end
# Option 2: Use Redis directly (recommended for scale)
Gitlab::Experiment.configure do |config|
config.cache = Gitlab::Experiment::Cache::RedisHashStore.new(
Redis.new(url: ENV['REDIS_URL']),
expires_in: 30.days
)
end
# Option 3: No caching (deterministic rollout strategies only)
config.cache = nil
The gem includes the RedisHashStore cache store, which is
documented in its implementation.
Important: Caching changes how rollout strategies behave. Once cached, subsequent calls return the cached value regardless of rollout strategy changes.
Advanced: Custom Rollout Strategies
Build custom integrations with your existing infrastructure:
Example: Flipper Integration
# We put it in this module namespace so we can get easy resolution when
# using `default_rollout :flipper` in our usage later.
module Gitlab::Experiment::Rollout
class Flipper < Percent
def enabled?
::Flipper.enabled?(experiment.name, self)
end
def flipper_id
"Experiment;#{id}"
end
end
end
So, Flipper needs something that responds to flipper_id, and since our experiment "id" (which is also our context key)
is unique and consistent, we're going to give that to Flipper to manage things like percentage of actors etc.
You might want to consider something more complex here if you're using things that can be flipper actors in your
experiment context.
Anyway, now you can use your custom Flipper rollout strategy by instantiating it in configuration:
Gitlab::Experiment.configure do |config|
config.default_rollout = Gitlab::Experiment::Rollout::Flipper.new
end
Or if you don't want to make that change globally, you can use it in specific experiment classes:
class PillColorExperiment < Gitlab::Experiment # OR ApplicationExperiment
# ...registered behaviors
default_rollout :flipper,
distribution: { control: 26, red: 37, blue: 37 } # optionally specify distribution
end
Now, enabling or disabling the Flipper feature flag will control if the experiment is enabled or not. If the experiment is enabled, as determined by our custom rollout strategy, the standard resolution logic will be executed, and a variant (or control) will be assigned.
experiment(:pill_color).enabled? # => false
experiment(:pill_color).assigned.name # => "control"
# Now we can enable the feature flag to enable the experiment.
Flipper.enable(:pill_color) # => true
experiment(:pill_color).enabled? # => true
experiment(:pill_color).assigned.name # => "red"
Middleware
There are times when you'll need to do link tracking in email templates, or markdown content -- or other places you won't be able to implement tracking. For these cases a middleware layer that can redirect to a given URL while also tracking that the URL was visited has been provided.
In Rails this middleware is mounted automatically, with a base path of what's been configured for mount_at.
If this path is nil, the middleware won't be mounted at all.
Gitlab::Experiment.configure do |config|
config.mount_at = '/experiment'
# Only redirect on permitted domains.
config.redirect_url_validator = ->(url) { (url = URI.parse(url)) && url.host == 'gitlab.com' }
end
Once configured to be mounted, the experiment tracking redirect URLs can be generated using the Rails route helpers.
ex = experiment(:example)
# Generating the path/url using the path and url helper.
experiment_redirect_path(ex, url: 'https//gitlab.com/docs') # => "/experiment/example:20d69a...?https//gitlab.com/docs"
experiment_redirect_url(ex, url: 'https//gitlab.com/docs') # => "https://gitlab.com/experiment/example:20d69a...?https//gitlab.com/docs"
# Manually generating a url is a bit less clean, but is possible.
"#{Gitlab::Experiment::Configuration.mount_at}/#{ex.to_param}?https//docs.gitlab.com/"
Testing (rspec support)
This gem comes with some rspec helpers and custom matchers. To get the experiment specific rspec support, require the rspec support file:
require 'gitlab/experiment/rspec'
Any file in spec/experiments path will automatically get the experiment specific support, but it can also be included
in other specs by adding the :experiment label:
describe MyExampleController do
context "with my experiment", :experiment do
# experiment helpers and matchers will be available here.
end
end
Stub helpers
You can stub experiment variant resolution using the stub_experiments helper. The helper supports multiple formats for
flexibility:
Simple hash format:
it "stubs experiments using hash format" do
stub_experiments(pill_color: :red)
experiment(:pill_color) do |e|
expect(e).to be_enabled
expect(e.assigned.name).to eq('red')
end
end
Hash format with options:
it "stubs experiments with assigned option" do
stub_experiments(pill_color: { variant: :red, assigned: true })
experiment(:pill_color) do |e|
expect(e).to be_enabled
expect(e.assigned.name).to eq('red')
end
end
Mixed formats (symbols and hashes together):
it "stubs multiple experiments with mixed formats" do
stub_experiments(
pill_color: :red,
hippy: { variant: :free_love, assigned: true },
yuppie: :financial_success
)
expect(experiment(:pill_color).assigned.name).to eq(:red)
expect(experiment(:hippy).assigned.name).to eq(:free_love)
expect(experiment(:yuppie).assigned.name).to eq(:financial_success)
end
Boolean true (allows rollout strategy to assign):
it "stubs experiments while allowing the rollout strategy to assign the variant" do
stub_experiments(pill_color: true) # only stubs enabled?
experiment(:pill_color) do |e|
expect(e).to be_enabled
# expect(e.assigned.name).to eq([whatever the rollout strategy assigns])
end
end
Testing only_assigned behavior
When you use the assigned: true option in stub_experiments, the find_variant method is automatically stubbed
to return the specified variant. This allows you to test the only_assigned behavior:
it "tests only_assigned behavior with a cached variant" do
stub_experiments(pill_color: { variant: :red, assigned: true })
experiment_instance = experiment(:pill_color, actor: user, only_assigned: true)
expect(experiment_instance).not_to be_excluded
expect(experiment_instance.run).to eq('red')
end
it "tests only_assigned behavior without a cached variant" do
stub_experiments(pill_color: :red)
experiment_instance = experiment(:pill_color, actor: user, only_assigned: true)
expect(experiment_instance).to be_excluded
expect(experiment_instance.run).to eq('red')
end
Note: The assigned: true option only works correctly when caching is disabled. When caching is enabled,
find_variant will attempt to read from the actual cache store rather than using the stub. In this case, you can
populate the cache naturally by running the experiment first to assign and cache a variant before testing with
only_assigned: true.
Registered behaviors matcher
It's useful to test our registered behaviors, as well as their return values when we implement anything complex in them.
The register_behavior matcher is useful for this.
it "tests our registered behaviors" do
expect(experiment(:pill_color)).to register_behavior(:control)
.with('grey') # with a default return value of "grey"
expect(experiment(:pill_color)).to register_behavior(:red)
expect(experiment(:pill_color)).to register_behavior(:blue)
end
Exclusion and segmentation matchers
You can also easily test your experiment classes using the exclude, segment matchers.
let(:excluded) { double(first_name: 'Richard', created_at: Time.current) }
let(:segmented) { double(first_name: 'Jeremy', created_at: 3.weeks.ago) }
it "tests the exclusion rules" do
expect(experiment(:pill_color)).to exclude(actor: excluded)
expect(experiment(:pill_color)).not_to exclude(actor: segmented)
end
it "tests the segmentation rules" do
expect(experiment(:pill_color)).to segment(actor: segmented)
.into(:red) # into a specific variant
expect(experiment(:pill_color)).not_to segment(actor: excluded)
end
Tracking matcher
Tracking events is a major aspect of experimentation, and because of this we try to provide a flexible way to ensure your tracking calls are covered.
before do
stub_experiments(pill_color: true) # stub the experiment so tracking is permitted
end
it "tests that we track an event on a specific instance" do
expect(subject = experiment(:pill_color)).to track(:clicked)
subject.track(:clicked)
end
You can use the on_next_instance chain method to specify that the tracking call could happen on the next instance of
the experiment.
This can be useful if you're calling experiment(:example).track downstream and don't have access to that instance.
Here's a full example of the methods that can be chained onto the track matcher:
it "tests that we track an event with specific details" do
expect(experiment(:pill_color)).to track(:clicked, value: 1, property: '_property_')
.on_next_instance # any time in the future
.with_context(foo: :bar) # with the expected context
.for(:red) # and assigned the correct variant
experiment(:pill_color, :red, foo: :bar).track(:clicked, value: 1, property: '_property_')
end
Tracking, anonymity and GDPR
We generally try not to track things like user identifying values in our experimentation. What we can and do track is the "experiment experience" (a.k.a. the context key).
We generate this key from the context passed to the experiment. This allows creating funnels without exposing any user information.
This library attempts to be non-user-centric, in that a context can contain things like a user or a project.
If you only include a user, that user would get the same experience across every project they view. If you only include the project, every user who views that project would get the same experience.
Each of these approaches could be desirable given the objectives of your experiment.
Development
After cloning the repo, run bundle install to install dependencies.
Running tests
The test suite requires Redis to be running.
Install and start Redis
(redis-server) before running tests.
Once Redis is running, execute the tests:
bundle exec rake
You can also run bundle exec pry for an interactive prompt that will allow you to experiment.
Contributing
Bug reports and merge requests are welcome on GitLab at https://gitlab.com/gitlab-org/ruby/gems/gitlab-experiment. This project is intended to be a safe, welcoming space for collaboration, and contributors are expected to adhere to the Contributor Covenant code of conduct.
Make sure to include a changelog entry in your commit message and read the changelog entries section.
Release process
Please refer to the Release Process.
License
The gem is available as open source under the terms of the MIT License.
Code of conduct
Everyone interacting in the Gitlab::Experiment project’s codebases, issue trackers, chat rooms and mailing lists is
expected to follow the code of conduct.