Optimizely will be sunsetting Full Stack Experimentation on July 29, 2024. See the recommended Feature Experimentation migration timeline and documentation.

Dev GuideAPI Reference
Dev GuideAPI ReferenceUser GuideLegal TermsGitHubDev CommunityOptimizely AcademySubmit a ticketLog In
Dev Guide

Event batching

How the Optimizely Python SDK uses the event processor to batch impressions and conversion events into a single payload before sending it to Optimizely.

The Optimizely Full Stack Python SDK lets you batch events and includes options to set a maximum batch size and flush interval timeout. The benefit of event batching means less network traffic for the same number of impression and conversion events tracked.

The event batching functionality works by processing impression events from Activate, Is Feature Enabled and conversion events from Track and placing them into a queue. The queue is drained when it either reaches its maximum size limit or when the flush interval is triggered.

By default, event batching is disabled in Python SDK versions 3.3.0 and newer.

📘

Note

Event batching works with both out-of-the-box and custom event dispatchers.

Make sure that you are not sending personally identifiable information (PII) to Optimizely. The event batching process does not remove PII from events.

Configure event batching

Event batching can be enabled through the usage of BatchEventProcessor. Optimizely provides two main options to configure event batching: batch_size and flush_interval. You can pass in both these options when creating instance of BatchEventProcessor and pass the created instance during Optimizely client creation. When using BatchEventProcessor, events are held in a queue until either:

  • The number of events reaches the defined batch_size.
  • The oldest event has been in the queue for longer than the defined flush_interval, which is specified in seconds. The queue is then flushed and all queued events are sent to Optimizely in a single network request.
  • A new datafile revision is received. This occurs only when live datafile updates are enabled.
from optimizely import optimizely
from optimizely import event_dispatcher as optimizely_event_dispatcher
from optimizely.event import event_processor

# Set event dispatcher that the 
# You can reference your own implementation of event dispatcher here
event_dispatcher = optimizely_event_dispatcher.EventDispatcher

# Create instance of BatchEventProcessor.
#
# In this example here we set batch size to 15 events
# and flush interval to 50 seconds. 
# Setting start_on_init starts the consumer
# thread to start receiving events.
#
# See table below for explanation of these
# and other configuration options.  
batch_processor = event_processor.BatchEventProcessor(
    event_dispatcher,
    batch_size=15,
    flush_interval=50
    start_on_init=True,
)

# Create Optimizely client and pass in instance 
# of BatchEventProcessor to enable batching.
optimizely_client = optimizely.Optimizely(
    sdk_key='<Your SDK Key here>', 
    datafile='<Your datafile here>',
    event_processor=batch_processor
)

The following table defines these and other options that you can use to configure the BatchEventProcessor.

NameDefault valueDescription
event_dispatcher No default value. Required parameter.An event handler to manage network calls.

Provides a dispatch_event method that takes in URL and parameters and dispatches request.
logger
optional
NoOpLoggerA logger implementation to log issues.
flush_interval
optional
30The maximum duration in seconds that an event can exist in the queue before being flushed.
batch_size
optional
10The maximum number of events to hold in the queue. Once this number is reached, all queued events are flushed and sent to Optimizely.
start_on_init
optional
FalseBoolean which if set true starts the thread consuming and queuing events on initializing the BatchEventProcessor.

By default the value is False and so the consumer thread is not ready to receive any events. One can always start the consumer thread by calling start on the instance of BatchEventProcessor
timeout_interval
optional
5Number representing time interval in seconds before joining the consumer thread.
event_queue
optional
six.moves.queue.QueueComponent that allows you to accumulate events until they are dispatched.
notification_center
optional
NotificationCenterInstance of optimizely.notification_center.NotificationCenter.

By default an instance of notification_center.NotificationCenter is created.

For more information, see Initialize SDK.

❗️

Warning

The maximum payload size is 3.5 MB. Optimizely rejects requests with a 400 response code, Bad Request Error, if the batch payload exceeds this limit.

The size limitation is because of the Optimizely Events API, which Full Stack uses to send data to Optimizely.

The most common cause of a large payload size is a high batch size. If your payloads exceed the size limit, try configuring a smaller batch size.

Side effects

The table lists other Optimizely functionality that may be triggered by using this class.

FunctionalityDescription
LogEventWhenever the event processor produces a batch of events, a LogEvent object will be created using the EventFactory.

It contains batch of conversion and impression events.

This object will be dispatched using the provided event dispatcher and also it will be sent to the notification subscribers.
Notification ListenersFlush invokes the LOG_EVENT notification listener if this listener is subscribed to.

Register LogEvent listener

To register a LogEvent notification listener:

def on_log_event():
  pass

optimizely_client.notification_center.add_notification_listener(enums.NotificationTypes.LOG_EVENT, on_log_event)

LogEvent

LogEvent object gets created using EventFactory. It represents the batch of impression and conversion events we send to the Optimizely backend.

ObjectTypeDescription
http_verb
Required (non null)
StringThe HTTP verb to use when dispatching the log event. It can be GET or POST.
url
Required (non null)
StringURL to dispatch log event to.
params
Required (non null)
DictEvent Batch. Contains all the information regarding every event which is batched. including list of visitors which contains UserEvent.
headersDictRequest Headers.

Stop BatchEventProcessor on application exit

If you enable event batching, it is important that you call the 'stop' method, batch_processor.stop(), prior to exiting. This ensures that queued events are flushed as soon as possible to avoid any data loss.

Important

Because the BatchEventProcessor maintains a buffer of queued events, you must call stop() on the BatchEventProcessor instance before shutting down your application.

MethodDescription
stop()Stops and flushes the event queue.

Note: Optimizely recommes that you connect this method to a kill signal for the running process.