Optimizely Agent is a standalone, open-source, and highly available microservice that provides major benefits over using Optimizely SDKs in certain use cases. The Agent REST API offers consolidated and simplified endpoints for accessing all the functionality of Optimizely Full Stack SDKs.
A typical production installation of Optimizely Agent is to run two or more services behind a load balancer or proxy. The service itself can be run via a Docker container or installed from source. See Setup Optimizely Agent for instructions on how to run Optimizely Agent.
Here are some of the top reasons to consider using Optimizely Agent:
If you already separate some of your logic into services that might need to access the Optimizely decision APIs, we recommend using Optimizely Agent.
The images below compare implementation styles in a service-oriented architecture, first without using Optimizely Agent, which shows six SDK embedded instances:
Now with Agent, instead of installing the SDK six times, you create just one Optimizely instance: an HTTP API that every service can access as needed.
If you want to deploy Optimizely Full Stack once, then roll out the single implementation across a large number of teams, we recommend using Optimizely Agent.
By standardizing your teams' access to the Optimizely service, you can better enforce processes and implement governance around feature management and experimentation as a practice.
You do not want many SDK instances connecting to Optimizely's cloud service from every node in your application. Optimizely Agent centralizes your network connection. Only one cluster of agent instances connects to Optimizely for tasks like update datafiles and dispatch events.
You are using a language that is not supported by a native SDK (i.e. Elixir, Scala, Perl). While it is possible to create your own service using an Optimizely SDK of your choice, you could also customize the open-source Optimizely Agent to your needs without building the service layer on your own.
If your use case would not benefit greatly from Optimizely Agent, you should consider the below reasons to not use Optimizely Agent and review Optimizely's many open-source SDKs instead.
If time to provide bucketing decisions is a primary concern for you, you may want to use an embedded Full Stack SDK rather than Optimizely Agent.
If your app is constructed as a monolith, embedded SDKs might be easier to install and might be a more natural fit for your application and development practices.
If you are looking for the fastest way to get a single team up and running with deploying feature management and experimentation, embedding an SDK is the best option for you at first. You can always start using Optimizely Agent later, and it can even be used alongside Optimizely Full Stack SDKs running in another part of your stack.
While every implementation is different, you can review this section of best practices for tips on these commonly discussed topics.
Agent can scale to large decision/event tracking volumes with relatively low CPU / memory specs. For example, at Optimizely, we scaled our deployment to 740 clients with a cluster of 12 agent instances, which in total use 6 vCPUs and 12GB RAM. You will likely need to focus more on network bandwidth than compute power.
Any standard load balancer should let you route traffic across your agent cluster. At Optimizely, we used an AWS Elastic Load Balancer (ELB) for our internal deployment. This allows us to transparently scale our agent cluster as internal demands increase.
Agent offers eventual rather than strong consistency across datafiles.
In detail, today, each agent instance maintains a dedicated, separate cache. Each agent instance persists an SDK instance for each SDK key your team uses. Agent instances automatically keep datafiles up to date for each SDK key instance so that you will have eventual consistency across the cluster. The rate of the datafile update can be set as the configuration value
OPTIMIZELY_CLIENT_POLLINGINTERVAL (the default is 1 minute).
Because SDKs are generally stateless today, they should not need to share data. We plan to add a common backing data store, so we invite you to share your feedback.
If you require strong consistency across datafiles, then we recommend an active/passive deployment where all requests are made to a single vertically scaled host, with a passive, standby cluster available for high availability.
Updated 4 months ago